Skip to main content

ORIGINAL RESEARCH article

Front. Genet., 07 June 2021
Sec. Computational Genomics

Automated Facial Recognition for Noonan Syndrome Using Novel Deep Convolutional Neural Network With Additive Angular Margin Loss

\r\nHang Yang,&#x;Hang Yang1,2†Xin-Rong Hu&#x;Xin-Rong Hu3†Ling SunLing Sun1Dian HongDian Hong1Ying-Yi ZhengYing-Yi Zheng4Ying XinYing Xin5Hui LiuHui Liu1Min-Yin Lin,Min-Yin Lin1,2Long WenLong Wen1Dong-Po LiangDong-Po Liang1Shu-Shui Wang*Shu-Shui Wang1*
  • 1Department of Pediatric Cardiology, Guangdong Provincial People’s Hospital, Guangdong Academy of Medical Sciences, Guangdong Cardiovascular Institute, Guangdong Provincial Key Laboratory of Structural Heart Disease, Guangzhou, China
  • 2Department of Pediatrics, Shantou University Medical College, Shantou, China
  • 3Department of Computer Science and Engineering, University of Notre Dame, South Bend, IN, United States
  • 4Cardiac Center, Guangdong Women and Children Hospital, Guangzhou, China
  • 5Department of Cardiology, Shenzhen Children’s Hospital, Shenzhen, China

Background: Noonan syndrome (NS), a genetically heterogeneous disorder, presents with hypertelorism, ptosis, dysplastic pulmonary valve stenosis, hypertrophic cardiomyopathy, and small stature. Early detection and assessment of NS are crucial to formulating an individualized treatment protocol. However, the diagnostic rate of pediatricians and pediatric cardiologists is limited. To overcome this challenge, we propose an automated facial recognition model to identify NS using a novel deep convolutional neural network (DCNN) with a loss function called additive angular margin loss (ArcFace).

Methods: The proposed automated facial recognition models were trained on dataset that included 127 NS patients, 163 healthy children, and 130 children with several other dysmorphic syndromes. The photo dataset contained only one frontal face image from each participant. A novel DCNN framework with ArcFace loss function (DCNN-Arcface model) was constructed. Two traditional machine learning models and a DCNN model with cross-entropy loss function (DCNN-CE model) were also constructed. Transfer learning and data augmentation were applied in the training process. The identification performance of facial recognition models was assessed by five-fold cross-validation. Comparison of the DCNN-Arcface model to two traditional machine learning models, the DCNN-CE model, and six physicians were performed.

Results: At distinguishing NS patients from healthy children, the DCNN-Arcface model achieved an accuracy of 0.9201 ± 0.0138 and an area under the receiver operator characteristic curve (AUC) of 0.9797 ± 0.0055. At distinguishing NS patients from children with several other genetic syndromes, it achieved an accuracy of 0.8171 ± 0.0074 and an AUC of 0.9274 ± 0.0062. In both cases, the DCNN-Arcface model outperformed the two traditional machine learning models, the DCNN-CE model, and six physicians.

Conclusion: This study shows that the proposed DCNN-Arcface model is a promising way to screen NS patients and can improve the NS diagnosis rate.

Introduction

Noonan syndrome (NS) is a genetically heterogeneous disorder with an estimated prevalence of 1 in 1,000–2,500, caused by germline mutations in 11 critical genes of the highly conserved Ras/Mitogen-Activated Protein Kinases (MAPK) pathway (Mendez and Opitz, 1985; Tajan et al., 2018). This multisystem disease is characterized by hypertelorism, ptosis, dysplastic pulmonary valve stenosis, hypertrophic cardiomyopathy, and small stature (Roberts et al., 2013; Li et al., 2019). Early detection and assessment of NS are crucial to formulating an individualized treatment protocol. NS can be diagnosed via clinical features and genetic testing (Van Der Burgt et al., 1994; Roberts et al., 2013). However, because of the complexity and rarity of NS, identifying it remains challenging for pediatric cardiologists and pediatricians. In the wake of this problem, an efficient and convenient auxiliary diagnostic approach is needed for the early diagnosis of NS.

Many genetic syndromes have craniofacial alterations (Hart and Hart, 2009), and facial appearance can be a momentous clue in making an early diagnosis of syndromes (Kuru et al., 2014). The utility of traditional machine learning methods and deep learning methods for diagnosing NS based on pattern recognition of face images has been explored previously by several researchers (Boehringer et al., 2006; Kruszka et al., 2017; Tekendo-Ngongang and Kruszka, 2020; Porras et al., 2021). In 2019, Gurovich et al. (2019) presented a deep DCNN framework, called DeepGestalt, trained on a database of over 17,000 pictures of faces representing more than 200 genetic syndromes. Gurovich et al. (2019) further applied the DeepGestalt model to discriminate five different genotypes of NS and predicted the five desired classes with a top-1 accuracy of 64%. However, to our best knowledge, no studies identified NS patients from healthy children and from children with several other genetic syndromes.

In the present study, therefore, we developed an automated facial recognition model for NS identification based on state-of-the-art tools in the field of facial recognition: a deep convolutional neural network (DCNN) and a novel loss function called Additive Angular Margin Loss (ArcFace; Deng et al., 2019). The main contributions of this study are the following: (1) to our knowledge, this is the first attempt at using a DCNN model with Arcface loss function to generate an automated facial recognition model (DCNN-Arcface model) to identify genetic syndromes; (2) the identification performance of the DCNN-Arcface model outranked two traditional machine learning models; (3) the identification performance of the DCNN-Arcface model was superior to the DCNN framework with cross-entropy loss function (the DCNN-CE model); (4) the identification performance of the DCNN-Arcface model outperformed physicians; and (5) the DCNN-Arcface model can distinguish NS patients from healthy children and from children with several other genetic syndromes.

Materials and Methods

Dataset

The dataset included 127 NS patients (68 males and 59 females), 163 healthy children, and 130 children with several other dysmorphic syndromes (see >Supplementary Table 1). The photo dataset contained only one frontal face image from each participant. Thirty-seven NS patients were recruited from Guangdong Provincial People’s Hospital, between January 2017 and September 2020. Other NS images were obtained from the medical literature (Xu et al., 2017; Leung et al., 2018; Li et al., 2019). The facial characteristics, demographic and genetic characteristics of the NS datasets are summarized in Figure 1 and Tables 1, 2.

FIGURE 1
www.frontiersin.org

Figure 1. Facial characteristics of NS patients collected from the Guangdong Provincial People’s Hospital, China (N = 37). The black bar is used to protect privacy.

TABLE 1
www.frontiersin.org

Table 1. Demographic and genetic characteristics of Noonan syndrome patients.

TABLE 2
www.frontiersin.org

Table 2. Pathogenic variants detected in Noonan syndrome patients collected from the Guangdong Provincial People’s Hospital.

All face images in this study were required to fulfill two eligibility criteria. First, the diagnosis of NS and other genetic syndromes was confirmed by fluorescence in situ hybridization, karyotype analysis or next-generation sequencing. Second, the faces should be sufficiently legible and oriented.

The DCNN-Arcface Model

The architecture of the DCNN-Arcface model is illustrated in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2. Architecture of DCNN-Arcface model for Noonan syndrome identification. We used convolutional layers with stride = 2 instead of max-pooling to half the feature map size and double channels number. After extracting the embeddings with multiple convolutional layers, we normalized the weights of the last fully connected layer ∥Wi∥ = 1 with L2 normalization and rescaled the norm of embedding vector to s, ∥χn∥ = s. Then, an angular margin penalty m was added to the target angle θl,m. After that, cos⁡(θl,m + m) was calculated, and all logits were multiplied by the feature scale s. The logits then went through the SoftMax function to derive the probability for each class. “DW conv” represents depth-wise convolution.

Image Preprocessing

The first step was to detect the patient’s face in an input image. Here, a multi-task convolutional neural network (MTCNN; Zhang et al., 2016) was applied to detect face image areas. It ultimately utilized five facial landmarks (Figure 3). The MTCNN contained an image pyramid and a three-stage cascaded framework. When a raw face image was given, different scale ratios were used to resize the face image to build an image pyramid. It was then transmitted to the three-stage cascaded framework as an input. After training on three convolutional networks [proposal network (P-Net), refine network (R-Net), and output network (O-Net)], the image pyramid was finally converted into five facial landmark positions. We subsequently cropped the face image containing the five facial landmarks into (3, 112, 112).

FIGURE 3
www.frontiersin.org

Figure 3. Pipeline of the multi-task convolutional neural network (MTCNN).

Feature Extraction and Image Embedding Using a Novel Deep Convolutional Neural Network

In this work, we used a novel two-dimension DCNN to extract face features and then embed the face images in N-dimensional vector space xn ∈ ℜN. The first layer was a traditional convolution block with a kernel size of 7 × 7, which was followed by multiple residual blocks (He et al., 2016) and depth-wise convolution blocks (Bai et al., 2018). Figure 4 depicts the construction of a single residual block and depth-wise (Dwise) convolution block. The Dwise convolution block consisted of three convolutional operations followed by a nonlinear unit, the Rectified Linear Unit (ReLU). The construction of the residual block was similar to the Dwise block, except for the addition of “shortcut connections” to feedforward neural networks. After multiple layers of feature extraction, we used a global average pooling layer to flatten the feature maps followed by a fully connected layer. The input face images were finally embedded in N-dimension vector space xn∈ℜN.

FIGURE 4
www.frontiersin.org

Figure 4. Illustration of a single depth-wise convolution block and residual block. (A) The construction of the Depth-wise (Dwise) convolution block denoting a convolutional layer with a convolution group number set as input channels. (B) The construction of the residual block. “Linear” means that there is no use of an activation function.

Arcface Loss Function

In this work, we used a loss function representing state-of-the-art in the field of facial recognition, called ArcFace loss function (Deng et al., 2019). The main idea is based on the observation that the weight matrix from the last fully connected layer can be viewed as a combination of vectors that represent the conceptual centers corresponding to different face classes. In detail, after extracting the embeddings with multiple convolutional layers, we rescaled the norm of the embedding vector to s, ∥χn∥ = s, where xnrepresents the embedding vector of the nth sample. Then, the individual wights of the last fully connected layer were normalized with L2 normalization, ∥Wi∥ = 1, in which Wi is the weight of the fully connected layer of class i. Thus, the product WiT*χn reflected the angle between the embedding vector and the weight of class i in the fully connected layer.

The loss function was then defined as,

L A = - n ω i log e W i T x n cos θ i , n j e W J T x n cos θ j , n (1)

where ωiis the weight of class i to deal with biased dataset. In our case, since our dataset has balanced of patient for each class, we set all ωi as 1.

By minimizing the loss function, embedding vectors for images within the same class were forced to gather around the same weight vector in the vector space, which enhanced intra-class compactness. Moreover, an additive angle margin m was added to improve the robustness and discrepancy for different classes, according to Deng et al. (2019).

L A = - n ω i log e s * cos ( θ i , n + m ) e s * cos ( θ i , n + m ) + j i e s * cos θ j , n (2)

Mathematically, the positive additive angle decreases the value of cos (θi,  n + m). If we try to maintain the value of the overall loss function, we can either reduce θi,  n or increase θj,  n. That is, we can bring the embedding vector closer to the weight vector of the same class or further away from other classes.

Setting

We first pre-trained our network on a public human face dataset, CASIA (Wen et al., 2016), and then retrained the whole network on the NS identification task. In the training process, we applied rotation and horizontal flipping to augment the dataset. Each face image was rotated at angles θ = [90°, 180°] and horizontally flipped. Thus, three face images were generated for each original face image. We used Adam as our optimizer, and the learning rate was set to 1e–4. Five-fold cross-validation was performed to evaluate the performance of DCNN-Arcface model, DCNN-CE model, and two traditional machine learning models. The proportion of training set, validation set, and test set was 3:1:1. For the DCCN-Arcface model, in the testing phase, we calculated the embeddings for all faces in the training set and derived the average embedding for each class as a reference vector. We used the angle between the inference image’s embedding vector and the reference vector for prediction instead of the weights on the last fully connected layer (Deng et al., 2019).

Comparative Experiments

In this study, we implemented a DCNN-CE model and two traditional machine learning models. Comparisons of the proposed DCNN-Arcface model with these three models were performed.

Comparison With Traditional Machine Learning Models

Before the prevalence of deep learning methods, various supervised machine learning models were used for NS identification (Boehringer et al., 2006; Kruszka et al., 2017). Therefore, we applied two supervised traditional machine learning methods to construct two additional facial recognition models for NS identification: a support vector machine (SVM) (SVM-Linear model) and logistic regression (LR) (LR model). These methods were unable to take image data as input directly. Therefore, the first step was to extract features from patients’ faces. We used a landmark detector from the “dlib” Python library to locate 68 informative points from each patient’s face, including the outline of the eyebrows, eyes, nose, mouth, and jaw. The landmark detector was an implementation of Kazemi and Sullivan (2014), in which an ensemble of regression trees is trained with manually labeled data to estimate the coordinates of facial landmarks. Then, two types of features—shape descriptor features and appearance descriptor features—were measured based on those landmarks. We calculated the distances between every two points as the shape information, although this was more complicated for the appearance descriptor feature. For each point, we derived the local binary pattern (LBP) with neighboring sample points set at 12 and the radius set at 4. Thus, we could compute six statistics for the LBP histogram: the mean, variance, skewness, kurtosis, energy, and entropy. We found that some landmarks were within a small area, such that features extracted from those points may provide redundant information for identification. Therefore, we selected 38 landmarks for feature measurement. The total number of features comprised 703 shape descriptor features and 266 appearance descriptor features. The concatenated feature vectors were then sent to the SVM with a linear kernel and to LR for NS identification. Feature selection was based on each feature’s weight in the classification model, and we treated smaller weights as being less significant to the identification task. We tried different numbers of features from 10 to 969. We found that the area under the ROC curve started converging with 400 for identifying NS patients from healthy children, and 300 features for identifying NS patients from children with other genetic syndromes. The detailed architecture is shown in Figure 5.

FIGURE 5
www.frontiersin.org

Figure 5. Illustration of two traditional machine learning models.

Comparison With the DCNN-CE Model

Cross-entropy (Rubinstein, 1999) loss function is the most widely used loss function for classification problems. It can provide a precise mathematical framework with high speed (Boer et al., 2005). Hence, we constructed another DCNN-based face recognition model of NS. The same DCNN framework were used, but Arcface loss function was substituted by the cross-entropy loss function.

Comparison With Physicians

In this study, we performed two experiments to determine the identification performance by six physicians with different areas of expertise (two pediatricians, two pediatric cardiologists, and two clinical geneticists). To distinguish NS patients from healthy children, the physicians were presented with a dataset consisting of all NS patients and healthy children in random order. They classified these face images as either NS patients or healthy children. Likewise, after rearranging all photographs of the NS patients and those of children with several other genetic syndromes, the physicians classified these face images as NS patients or children with several other genetic syndromes. Each face image was shown for 10 s without exhibiting any clinical data. The experiments were repeated three times.

Evaluation Metric

The metrics we used to evaluate the identification performance in all comparative experiments were the total identification accuracy measure, sensitivity measure, and specificity measure. Moreover, the area under the receiver operating characteristic curve (AUC) and the area under the precision–recall curve (AP scores) were also determined. These measures were calculated as follows:

T o t a l A c c u r a c y = T P + T N A m o u n t o f all images
S e n s i t i v i t y = T P T P + F N
S p e c i f i c i t y = T N T N + F P

where TP, TN, FP, and FN denote true positives, true negatives, false positives, and false negatives, respectively.

All measurements are reported as the mean ± standard deviation. These measurements indicated the ability of all methods to correctly distinguish NS patients from healthy children and children with several other genetic syndromes.

Statistical Analysis

McNemar’s test was used to determine the disagreement for binary outputs between DCNN-Arcface model and DCNN-CE model, DCNN-Arcface model and traditional machine learning models (Dietterich, 1998). Z-tests were constructed to compare the AUC and AP scores of DCNN-Arcface model, DCNN-CE model, and two machine learning models (Zhang et al., 2002). p-values < 0.05 were considered statistically significant.

Results

Accuracy, Specificity, Sensitivity, AUC, and AP Score of Different Models

Tables 3, 4 list the accuracy, specificity, sensitivity, AUC, and AP score of the proposed DCNN-Arcface model, DCNN-CE model, SVM-linear model, and LR model.

TABLE 3
www.frontiersin.org

Table 3. The accuracy, specificity, sensitivity, AUC, and AP score of different models at distinguishing Noonan syndrome from healthy children.

TABLE 4
www.frontiersin.org

Table 4. The accuracy, specificity, sensitivity, AUC, and AP score of different models at distinguishing Noonan syndrome from patients with several other genetic syndromes.

Comparison With Traditional Machine Learning Models

At distinguishing NS patients from healthy children, the DCNN-Arcface model achieved the best identification performance compared with the SVM-Linear model (p = 0.0002, McNemar’s test) and with the LR model (p = 0.0001, McNemar’s test). At distinguishing NS patients from children with several other genetic syndromes, the performance of the DCNN-Arcface model was also superior to the SVM-Linear model (p = 0.0001, McNemar’s test) and the LR model (p = 0.0000, McNemar’s test). The ROC curves and precision–recall (P–R) curves of different models with different tasks are shown in Figures 6, 7. The highest AUC and AP scores were achieved by the DCNN-Arcface model on both tasks (p = 0.0000, z-test). The ROC curves and P–R curves shown in Figures 6, 7 also indicate that the DCNN-Arcface model can significantly improve the identification performance of an NS facial recognition model.

FIGURE 6
www.frontiersin.org

Figure 6. ROC curves and P–R curves of four different models when distinguishing children with Noonan syndrome from healthy children. The DCNN-Arcface model is consistently better than the other three.

FIGURE 7
www.frontiersin.org

Figure 7. ROC curves and P–R curves of four different models when distinguishing children with Noonan syndrome from those with several other genetic syndromes. The DCNN-Arcface model is consistently better than the other three.

Comparion With the DCNN-CE Model

At distinguishing NS patients from healthy children, the performance of the DCNN-Arcface model outranked the DCNN-CE model (p = 0.0016, McNemar’s test). At distinguishing NS patients from children with several other genetic syndromes, better performance was also obtained from the DCNN-Arcface model (p = 0.0045, McNemar’s test). Significantly statistical difference was achieved when compare the AUC and AP score obtained from the DCNN-Arcface model with that of the DCNN-CE model (p = 0.0000, z-test). The ROC curves and P–R curves shown in Figures 6, 7 also elucidated the improved identification performance of NS facial recognition by applying the DCNN-Arcface model.

Comparison With Physicians

The identification performance of different physicians for different tasks is presented in Tables 5, 6. All six physicians completed the identification tasks, with an average accuracy of 0.7595 ± 0.0618 for identifying NS patients from healthy children, and 0.5826 ± 0.0303 for identifying NS patients from children with several other genetic syndromes. When classifying the physicians depend on their expertise, the clinical geneticists exhibited the best identification performance among the physicians, with the average accuracy of 0.7685 ± 0.1223 for identifying NS patients from healthy children, and 0.6105 ± 0.0106 for identifying NS patients from children with several other genetic syndromes. However, the DCNN-Arcface model outperformed all six physicians according to all metrics on both tasks, including accuracy, specificity, and sensitivity (Tables 3, 4).

TABLE 5
www.frontiersin.org

Table 5. Identification performance of different physicians at distinguishing Noonan syndrome patients from healthy children.

TABLE 6
www.frontiersin.org

Table 6. Identification performance of different physicians at distinguishing Noonan syndrome patients from children with several other genetic syndromes.

Discussion

Distinctive facial appearance provides significant information for predicting a specific genetic syndrome (Roosenboom et al., 2016). NS’s facial gestalt usually includes a high forehead, low posterior hairline, hypertelorism, highly arched palate, downslanting palpebral fissures, epicantal folds, ptosis, high wide peaks of the vermilion, deeply grooved philtrum, and low-set and posteriorly rotated ears (Tartaglia et al., 2011; Li et al., 2019). The diagnosis of NS could start with astute clinicians recognizing the specific facial dysmorphism. As they age, however, the faces of children with NS can become more atypical, as the face lengthens and becomes more triangular in shape (Allanson, 2016). Some genotypes of NS present with atypical facial characteristics (Jenkins et al., 2020). In addition, the facial appearance of NS is similar to that of other RASopathies, such as Cardio-faciocutaneous and Costello syndromes (Allanson, 2016). In this context, discriminating this particular syndrome based on facial appearance is challenging, and can lead to misdiagnose and misclassification.

In 2003, Loos et al. (2003) first used Gabor wavelet transformation, a machine learning method, to classify five syndromes with an accuracy of 76%. Machine learning is a subfield of Artificial Intelligence that allows computers to learn from data to make predictions for a given task without being explicitly programmed (Nguyen et al., 2019). Since then, other traditional machine learning methods have been reported for NS facial recognition. In 2006, Boehringer et al. (2006) first used principal component analysis to reduce covariates from face images of NS patients, and then applied linear discriminant analysis, SVM and k-th nearest neighbors to discriminate NS from other genetic syndromes with a maximum accuracy of 79.4%. In 2017, Kruszka et al. (2017) applied the independent component analysis to locate facial landmarks, and subsequently used a linear support vector machine as a classifier to classify NS patients and healthy children. They achieved sensitivites and specificities for Caucasian, African, Asian, and Latin American children of 0.95 and 0.93, 0.94 and 0.91, 0.95 and 0.90, and 0.96 and 0.98, respectively (Kruszka et al., 2017). Recently, Porras et al. (2021) applied the method presented by Kruszka et al. (2017) to discriminate patients with NS from those with Williams-Beuren syndromes, and obtained an accuracy of 85.68%. In the above methods, after extracting features with traditional machine learning methods, feature selection is usually performed to accelerate the relevancy and prevent over-fitting. However, both of these steps are complicated and time-consuming (Wang et al., 2018). With this limitation, the computer vision community has shifted toward DCNNs for medical image classification tasks (Zhang et al., 2019).

A DCNN is a feed-forward artificial neural network that consists of multiple convolutional layers followed by a nonlinear unit (Ongsulee, 2017). It can automatically learn representation from labeled data without requiring human expertise for feature extraction (Wang et al., 2018). Complicated layers are often constructed to achieve more satisfactory accuracy in facial recognition tasks (He et al., 2016). However, accuracy may be saturated and it can degrade rapidly as the network becomes deeper. In 2016, He et al. (2016) introduced a deep residual learning framework to solve this degradation problem by adding a residual block. The residual block mainly involved a “shortcut connection” that denotes the outputs of the identity mapping added to the outputs of the stacked layers. With this residual block, complicated neural networks are more easily trained (He et al., 2016). Nevertheless, the deep residual learning framework is still computationally expensiveness with respect to its size and content. A depth-wise convolution block can lower this computational complexity. Depth-wise convolution splits the standard convolution into two separate layers for filtering and combining. Through this optimization, the convolutional operation becomes more efficient in terms of the number of parameters and the computational cost (Bai et al., 2018). In the present study, we combined the residual block and the depth-wise block to develop a novel DCNN framework. The joint combination of these two blocks enabled us to build a light architecture without sacrificing accuracy.

The loss function guides the DCNN to extract features from input images by backpropagating gradients to the weights in the network. The definition of a loss function can impact the discriminative ability of a model. Cross-entropy is the most widely used loss function for classification tasks. However, cross-entropy fails to teach the neural network the similarity among samples belonging to the same class (Deng et al., 2019). The ArcFace is a novel loss function developed by Deng et al. (2019). By adding an additive angular margin penalty between deep features and the ground truth weight, Arcface loss function can simultaneously enhance the intra-class compactness compared to cross-entropy (Deng et al., 2019). Chinapas et al. (2019) adapted Arcface loss function to train personal verification system and achieved a maximum accuracy of 0.996. However, to the best of our knowledge, this loss function has not been used for genetic syndrome facial recognition tasks. NS has 12 different genotypes, and some of the genotypes present with an atypical facial appearance (Allanson et al., 2010). Due to this intrinsic intra-class variation, learning the discriminative features of NS is challenging. Arcface loss function addresses this problem with its distinctive ability for compact intra-class variation. Hence, it is more suitable for NS identification.

In the present study, we implemented the DCNN framework and Arcface loss function to construct an automated facial recognition model for NS identification. The DCNN-Arcface model achieved an accuracy of 0.9201 ± 0.0138 and an AUC of 0.9797 ± 0.0055 when distinguishing NS patients from healthy children, and an accuracy of 0.8171 ± 0.0074 and an AUC of 0.9274 ± 0.0062 when distinguishing NS patients from children with several other genetic syndromes. It outperformed all six physicians in terms of accuracy, sensitivity, and specificity. NS usually presents with considerable heterogeneity in clinical manifestations (Roberts et al., 2013), and it is a rare syndrome. As such, prompt diagnosis of NS in routine clinical practice is still a cumbersome problem for physicians. Although previous literature has shown that the DCNN-based facial recognition models can assist in diagnosing genetic syndromes (Gurovich et al., 2019; Qin et al., 2020), only a few studies have used DCNNs to identify NS. In 2020, Tekendo-Ngongang and Kruszka (2020) applied DeepGestalt, a DCNN-based architecture, to develop a NS facial recognition model. Their model discriminated NS patients from matched healthy individuals with an AUC of 0.979. However, the DeepGestalt model used cross-entropy as a loss function. By using a novel DCNN framework and Arcface, our DCNN-Arcface model can efficiently discriminate NS children from both healthy children and children with several other genetic syndromes. Also, the DCNN-Arcface model is more suitable for identifying NS. Our study offers compelling evidence that the DCNN-Arcface model can improve the diagnostic rate of NS. Our results also indicate that the DCNN-Arcface model can be adapted to detect other heterogeneous genetic syndromes.

The DCNN-Arcface model also outperformed two traditional machine learning methods. The AUC of DCNN-Arcface model was 0.9797 ± 0.0055 when discriminating NS patients from healthy children, while the AUCs of the two traditional machine learning models in the same task were 0.9031 ± 0.0064 and 0.8669 ± 0.0035, respectively. There are several possible explanations for this result. First, the DCNN-based model has many more parameters than the machine learning-based models, leading to better representation ability for fitting into the unknown function of input images and output prediction. The deep structure also enables the network to extract latent features layer-by-layer from raw images of NS patients’ faces. Moreover, not all selected features are informative for NS identification with the traditional machine learning method, and other useful features may be lost. In contrast, the DCNN performs feature extraction and classification in an end-to-end manner, which avoids any manual feature-selection bias. Finally, the ArcFace loss function increases the neural network’s discriminative power for different classes, while the loss function for the traditional machine learning model does not provide this benefit (Wang et al., 2018; Deng et al., 2019; Zhang et al., 2020).

The primary limitation of this study is that there were a limited number of dysmorphic facial photographs of NS patients. This might have led to over-fitting. In the future, we will conduct a multicenter study to collect more photographs, and we will explore the use of data augmentation methods, such as the generative adversarial networks, to generate more face images of NS patients.

Conclusion

In conclusion, this study illustrated that the proposed facial recognition model based on DCNN and Arcface loss function could play a prominent role in NS diagnosis. The results highlight the feasibility of facial recognition technology to identify NS in clinical practice.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

The studies involving human participants were reviewed and approved by Research Ethics Committee of Guangdong Provincial People’s Hospital. Written informed consent to participate in this study was provided by the participants’ legal guardian/next of kin. Written informed consent was obtained from the minor(s)’ legal guardian/next of kin for the publication of any potentially identifiable images or data included in this article.

Author Contributions

S-SW: conceptualization, funding acquisition, project administration, supervision, writing – review and editing, and funding. HY and X-RH: data curation, methodology, validation, writing – original draft. HY, X-RH, LS, DH, Y-YZ, YX, HL, M-YL, LW, and D-PL: resources. HY: software. All authors contributed to the article and approved the submitted version.

Funding

This research was funded by the National Natural Science Foundation of China (grant number 82070321), the Shenzhen San-ming Project, and the Guangdong Peak Project of Guangdong Province, China (grant number DFJH201802).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The reviewer, YS, declared a shared affiliation with one of the author, X-RH, to the handling editor at the time of review.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fgene.2021.669841/full#supplementary-material

Supplementary Table | Characterization of the mixed dysmorphic genetic syndromes dataset.

References

Allanson, J. E. (2016). Objective studies of the face of Noonan, Cardio-facio-cutaneous, and Costello syndromes: a comparison of three disorders of the Ras/MAPK signaling pathway. Am. J. Med. Genet. A 170, 2570–2577. doi: 10.1002/ajmg.a.37736

PubMed Abstract | CrossRef Full Text | Google Scholar

Allanson, J. E., Bohring, A., Dörr, H. G., Dufke, A., Gillessen Kaesbach, G., Horn, D., et al. (2010). The face of Noonan syndrome: does phenotype predict genotype? Am. J. Med. Genet. A 152A, 1960–1966. doi: 10.1002/ajmg.a.33518

PubMed Abstract | CrossRef Full Text | Google Scholar

Bai, L., Zhao, Y. M., and Huang, X. M. (2018). A CNN accelerator on FPGA using depthwise separable convolution. IEEE. Trans. Circuits II. 65, 1415–1419. doi: 10.1109/TCSII.2018.2865896

CrossRef Full Text | Google Scholar

Boehringer, S., Vollmar, T., Tasse, C., Wurtz, R. P., Gillessen-Kaesbach, G., Horsthemke, B., et al. (2006). Syndrome identification based on 2D analysis software. Eur. J. Hum. Genet. 14, 1082–1089. doi: 10.1038/sj.ejhg.5201673

PubMed Abstract | CrossRef Full Text | Google Scholar

Boer, P. T. D., Krosese, D. P., Mannor, S., and Rubinstein, R. Y. (2005). A tutorial on the cross-entropy method. Ann. Oper. Res. 134, 19–67. doi: 10.1007/s10479-005-5724-z

CrossRef Full Text | Google Scholar

Chinapas, A., Polpinit, P., Intiruk, N., and Saikaew, K. R. (2019). Personal verification system using ID card and face photo. Int. J. Mach. Learn. Comput. 9, 407–412. doi: 10.18178/ijmlc.2019.9.4.818

CrossRef Full Text | Google Scholar

Deng, J., Guo, J., Xue, N., and Zafeiriou, S. (2019). “ArcFace: Additive Angular Margin Loss for Deep Face Recognition,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Los Angeles, CA, 4690–4699. doi: 10.1109/CVPR.2019.00482

CrossRef Full Text | Google Scholar

Dietterich, T. (1998). Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput. 10, 1895–1923. doi: 10.1162/089976698300017197

PubMed Abstract | CrossRef Full Text | Google Scholar

Gurovich, Y., Hanani, Y., Bar, O., Nadav, G., Fleischer, N., Gelbman, D., et al. (2019). Identifying facial phenotypes of genetic disorders using deep learning. Nat. Med. 25, 60–64. doi: 10.1038/s41591-018-0279-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Hart, T. C., and Hart, P. S. (2009). Genetic studies of craniofacial anomalies: clinical implications and applications. Orthod. Craniofac. Res. 12, 212–220. doi: 10.1111/j.1601-6343.2009.01455.x

PubMed Abstract | CrossRef Full Text | Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep Residual Learning for Image Recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NH, doi: 10.1109/CVPR.2016.90

CrossRef Full Text | Google Scholar

Jenkins, J., Barnes, A., Birnbaum, B., Papagiannis, J., Thiffault, I., and Saunders, C. J. (2020). LZTR1-related hypertrophic cardiomyopathy without typical noonan syndrome features. Circ. Genom. Precis. Med. 13:e002690.

Google Scholar

Kazemi, V., and Sullivan, J. (2014). “One Millisecond Face Alignment with an Ensemble of Regression Trees,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, 1867–1874. doi: 10.1109/cvpr.2014.241

CrossRef Full Text | Google Scholar

Kruszka, P., Porras, A. R., Addissie, Y. A., Moresco, A., Medrano, S., and Mok, G. T. K. (2017). Noonan syndrome in diverse populations. Am. J. Med. Genet. A 173, 2323–2334. doi: 10.1002/ajmg.a.38362

PubMed Abstract | CrossRef Full Text | Google Scholar

Kuru, K., Niranjan, M., Tunca, Y., Osvank, E., and Azim, T. (2014). Biomedical visual data analysis to build an intelligent diagnostic decision support system in medical genetics. Artif. Intell. Med. 62, 105–118.

Google Scholar

Leung, G. K. C., Luk, H. M., Tang, V. H. M., Gao, W. W., Mak, C. C. Y., and Yu, M. H. C. (2018). Integrating functional analysis in the next-generation sequencing diagnostic pipeline of RASopathies. Sci. Rep. 8:2421. doi: 10.1038/s41598-018-20894-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, X., Yao, R., Tan, X., Li, N., Ding, Y., Li, J., et al. (2019). Molecular and phenotypic spectrum of Noonan syndrome in Chinese patients. Clin. Genet. 96, 290–299. doi: 10.1111/cge.13588

PubMed Abstract | CrossRef Full Text | Google Scholar

Loos, H. S., Wieczorek, D., Würtz, R. P., Malsburg, C. V. D., and Horsthemke, B. (2003). Computer-based recognition of dysmorphic faces. Eur. J. Hum. Genet. 11, 555–560. doi: 10.1038/sj.ejhg.5200997

PubMed Abstract | CrossRef Full Text | Google Scholar

Mendez, H., and Opitz, J. (1985). Noonan syndrome: a review. Am. J. Med. Genet. 21, 493–506.

Google Scholar

Nguyen, G., Dlugolinsky, S., Bobák, M., Tran, V., López García, Á, Heredia, I., et al. (2019). Machine learning and deep learning frameworks and libraries for large-scale data mining: a survey. Artif. Intell. Rev. 52, 77–124. doi: 10.1007/s10462-018-09679-z

CrossRef Full Text | Google Scholar

Ongsulee, P. (2017). “Artificial Intelligence, Machine Learning and Deep Learning,” in Proceedings of the 15th International Conference on ICT and Knowledge Engineering (ICT&KE), Bangkok, 1–6. doi: 10.1109/ICTKE.2017.8259629

CrossRef Full Text | Google Scholar

Porras, A., Summar, M., and Linguraru, G. (2021). Objective differential diagnosis of Noonan and Williams-Beuren syndromes in diverse populations using quantitative facial phenotyping. Mol. Genet. Genomic. Med. e1636. doi: 10.1002/mgg3.1636

CrossRef Full Text [Epub ahead of print], | PubMed Abstract | Google Scholar

Qin, B., Liang, L., Wu, J., Quan, Q., Wang, Z., and Li, D. (2020). Automatic identification of down syndrome using facial images with deep convolutional neural network. Diagnostics 10:487. doi: 10.3390/diagnostics10070487

PubMed Abstract | CrossRef Full Text | Google Scholar

Roberts, A. E., Allanson, J. E., Tartaglia, M., and Gelb, B. D. (2013). Noonan syndrome. Lancet 381, 333–342. doi: 10.1016/S0140-6736(12)61023-X

CrossRef Full Text | Google Scholar

Roosenboom, J., Hens, G., Mattern, B. C., Shriver, M. D., and Claes, P. (2016). Exploring the underlying genetics of craniofacial morphology through various sources of knowledge. Biomed. Res. Int. 2016:3054578. doi: 10.1155/2016/3054578

PubMed Abstract | CrossRef Full Text | Google Scholar

Rubinstein, R. (1999). The cross-entropy method for combinatorial and continuous optimization. Methodol. Comput. Appl. 1, 127–190. doi: 10.1023/A:1010091220143

CrossRef Full Text | Google Scholar

Tajan, M., Paccoud, R., Branka, S., Edouard, T., and Yart, A. (2018). The RASopathy family: consequences of germline activation of the RAS/MAPK pathway. Endocr. Rev. 39, 676–700. doi: 10.1210/er.2017-00232

PubMed Abstract | CrossRef Full Text | Google Scholar

Tartaglia, M., Gelb, B. D., and Zenker, M. (2011). Noonan syndrome and clinically related disorders. Best Pract. Res. Clin. Endocrinol. Metab. 25, 161–179. doi: 10.1016/j.beem.2010.09.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Tekendo-Ngongang, C., and Kruszka, P. (2020). Noonan syndrome on the African continent. Birth Defects Res. 112, 718–724. doi: 10.1002/bdr2.1675

PubMed Abstract | CrossRef Full Text | Google Scholar

Van Der Burgt, I., Berends, E., Lommen, E., Van Beersum, S., Hamel, B., and Mariman, E. (1994). Clinical and molecular studies in a large Dutch family with Noonan syndrome. Am. J. Med. Genet. 53, 187–191. doi: 10.1002/ajmg.1320530213

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, J., Ma, Y., Zhang, L., Gao, R. X., and Wu, D. (2018). Deep learning for smart manufacturing:methods and applications. J. Manuf. Syst. 48, 144–156. doi: 10.1016/j.jmsy.2018.01.003

CrossRef Full Text | Google Scholar

Wen, Y. D., Zhang, K. P., Li, Z. F., and Qiao, Y. (2016). “A Discriminative Feature Learning Approach for Deep Face Recognition,” in Proceedings of the Computer vision-European Conference on Computer Vision 2016, Amsterdam, 499–515. doi: 10.1007/978-3-319-46478-7_31

CrossRef Full Text | Google Scholar

Xu, S., Fan, Y., Sun, Y., Wang, L., Gu, X., and Yu, Y. (2017). Targeted/exome sequencing identified mutations in ten Chinese patients diagnosed with Noonan syndrome and related disorders. BMC Med. Genomics 10:62. doi: 10.1186/s12920-017-0298-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, D. D., Zhou, X. H., Freeman, H.Jr., and Freeman, J. L. (2002). A non-parametric method for the comparison of partial areas under ROC curves and its application to large health care data sets. Statist. Med. 21, 701–715. doi: 10.1002/sim1011

CrossRef Full Text | Google Scholar

Zhang, J., Xie, Y., Wu, Q., and Xia, Y. (2019). Medical image classification using synergic deep learning. Med. Image Anal. 54, 10–19. doi: 10.1016/j.media

CrossRef Full Text | Google Scholar

Zhang, K., Zhang, Z., Li, Z., and Qiao, Y. (2016). Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Proc. Lett. 23, 1499–1503. doi: 10.1109/LSP.2016.2603342

CrossRef Full Text | Google Scholar

Zhang, S., Zhang, S., Wang, B., and Habetler, T. G. (2020). Deep learning algorithms for bearing fault diagnostics–a comprehensive review. IEEE Access 8, 29857–29881. doi: 10.1109/ACCESS.2020.2972859

CrossRef Full Text | Google Scholar

Keywords: noonan syndrome, facial recognition model, deep learning, Arcface loss function, genetic syndromes

Citation: Yang H, Hu X-R, Sun L, Hong D, Zheng Y-Y, Xin Y, Liu H, Lin M-Y, Wen L, Liang D-P and Wang S-S (2021) Automated Facial Recognition for Noonan Syndrome Using Novel Deep Convolutional Neural Network With Additive Angular Margin Loss. Front. Genet. 12:669841. doi: 10.3389/fgene.2021.669841

Received: 19 February 2021; Accepted: 12 May 2021;
Published: 07 June 2021.

Edited by:

Liang Zhao, Taihe Hospital, Hubei University of Medicine, China

Reviewed by:

Yiyu Shi, University of Notre Dame, United States
Antonio R. Porras, University of Colorado Anschutz Medical Campus, United States
Fu Lijun, Shanghai Children’s Medical Center, China

Copyright © 2021 Yang, Hu, Sun, Hong, Zheng, Xin, Liu, Lin, Wen, Liang and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Shu-Shui Wang, wsscome@126.com

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.