ORIGINAL RESEARCH article

Front. Comput. Sci., 24 March 2021

Sec. Human-Media Interaction

Volume 3 - 2021 | https://doi.org/10.3389/fcomp.2021.624558

Recognition of Alzheimer’s Dementia From the Transcriptions of Spontaneous Speech Using fastText and CNN Models

  • MILE Laboratory, Department of Electrical Engineering, Indian Institute of Science, Bengaluru, India

Abstract

Alzheimer’s dementia (AD) is a type of neurodegenerative disease that is associated with a decline in memory. However, speech and language impairments are also common in Alzheimer’s dementia patients. This work is an extension of our previous work, where we had used spontaneous speech for Alzheimer’s dementia recognition employing log-Mel spectrogram and Mel-frequency cepstral coefficients (MFCC) as inputs to deep neural networks (DNN). In this work, we explore the transcriptions of spontaneous speech for dementia recognition and compare the results with several baseline results. We explore two models for dementia recognition: 1) fastText and 2) convolutional neural network (CNN) with a single convolutional layer, to capture the n-gram-based linguistic information from the input sentence. The fastText model uses a bag of bigrams and trigrams along with the input text to capture the local word orderings. In the CNN-based model, we try to capture different n-grams (we use n = 2, 3, 4, 5) present in the text by adapting the kernel sizes to n. In both fastText and CNN architectures, the word embeddings are initialized using pretrained GloVe vectors. We use bagging of 21 models in each of these architectures to arrive at the final model using which the performance on the test data is assessed. The best accuracies achieved with CNN and fastText models on the text data are 79.16 and 83.33%, respectively. The best root mean square errors (RMSE) on the prediction of mini-mental state examination (MMSE) score are 4.38 and 4.28 for CNN and fastText, respectively. The results suggest that the n-gram-based features are worth pursuing, for the task of AD detection. fastText models have competitive results when compared to several baseline methods. Also, fastText models are shallow in nature and have the advantage of being faster in training and evaluation, by several orders of magnitude, compared to deep models.

1 Introduction

Dementia is a syndrome characterized by the decline in cognition that is significant enough to interfere with one’s independent, daily functioning. Alzheimer’s disease contributes to around 60–70% of dementia cases. Toward the final stages of Alzheimer’s dementia (AD), the patients lose control of their physical functions and depend on others for care. As there are no curative treatments for dementia, the early detection is critical to delay or slow down the onset or progression of the disease. The mini-mental state examination (MMSE) is a widely used test to screen for dementia and to estimate the severity and progression of cognitive impairment.

AD affects the temporal characteristics of spontaneous speech. Changes in the spoken language are evident even in mild AD patients. Subtle language impairments such as difficulties in word finding and comprehension, usage of incorrect words, ambiguous referents, loss of verbal fluency, speaking too much at inappropriate times, talking too loudly, repeating ideas, and digressing from the topic are common in the early stages of AD (Savundranayagam et al., 2005) and they turn extreme in the moderate and severe stages. Szatlóczki et al. (2015) show that AD can be detected with the help of a linguistic analysis more sensitively than with other cognitive examinations. Mueller et al. (2018b) analyzed the connected language samples obtained from simple picture description tasks and found that the speech fluency and the semantic content features declined faster in participants with early mild cognitive impairment. The language profile of AD patients is characterized by “empty speech,” devoid of content words (Nicholas et al., 1985). They tend to use pronouns without proper noun references and indefinite terms like “this,” “that,” and “thing” more often (Mueller et al., 2018a). These results motivate us to believe that modeling the transcriptions of the narrative speech in the cookie-theft picture description task using n-gram language models can help in the detection of AD and prediction of MMSE score.

In this work we address the AD detection and MMSE score prediction problems using two natural language processing (NLP)–based models: 1) fastText and 2) convolutional neural network (CNN). These models have the advantage that they can be easily structured to capture the linguistic cues in the form of n-grams from the transcriptions of the picture description task, provided with the Alzheimer’s Dementia Recognition through Spontaneous Speech (ADReSS) dataset (Luz et al., 2020). CNNs, though originated in computer vision, have become popular for NLP tasks and have achieved great results in sentence classification (Kim, 2014), semantic parsing (tau Yih et al., 2014), search query retrieval (Shen et al., 2014), and other traditional NLP tasks (Collober et al., 2011). Our convolutional neural network model draws inspiration from the work on sentence classification using CNNs (Kim, 2014). The fastText (Joulin et al., 2017) is a simple and efficient model for text classification (e.g., tag prediction and sentiment analysis). The fundamental idea in the fastText classifier is to calculate the n-grams of an input sentence and append them to the end of the sentence. Our choice of fastText model is also motivated by its ability to often outperform deep learning classifiers in terms of accuracy and training/evaluation times (Joulin et al., 2017).

The rest of the paper is organized as follows. Section 2 discusses the ADReSS dataset in detail. Section 3 discusses the baseline results in AD detection. Section 4 discusses our proposed NLP-based models followed by the listing of results in Section 5. Our results and conclusions are discussed in Section 6.

2 ADReSS Dataset

The ADReSS dataset (Luz et al., 2020) is designed to provide Alzheimer’s research community with a standard platform for AD detection and MMSE score prediction. The dataset is acoustically preprocessed and balanced in terms of age and gender. It consists of audio recordings and transcriptions [in CHAT format (Macwhinney, 2009)] of the cookie-theft picture description task, elicited from subjects in the age group of 50–80 years. The training set consists of data from 108 subjects, 54 each from AD and non-AD classes. The test set has data from 48 subjects, again balanced with respect to AD and non-AD classes. More information on the ADReSS dataset can be found in the ADReSS challenge baseline paper (Luz et al., 2020).

3 Review of Baseline Methods

This section provides a brief overview of the various approaches for AD detection and MMSE score prediction on ADReSS dataset. These approaches can be broadly classified into three types based on the type of the features used in the problem: 1) acoustic feature, 2) linguistic feature, and 3) a fusion of acoustic and linguistic features. The performance of different approaches on the AD detection and MMSE score prediction tasks are compared using the accuracy and root mean square error (RMSE) measures computed on the ADReSS test set.where N is the total number of subjects involved in the study, the number of true positives, and the number of true negatives. and are the estimated and target MMSE scores for test sample. The results of different approaches on the ADReSS dataset are summarized in Table 1.

TABLE 1

ModelAccuracy (%)RMSE
Searle et al. (2020), DistilBERT81.254.58
Searle et al. (2020), SVM + CRF81.255.22
Pompili et al. (2020), x-vectors SRE54.17
Pompili et al. (2020), sentence embedding72.92
Pompili et al. (2020), fusion of system81.25
Luz et al. (2020), linguistic75.005.20
Sarawgi et al. (2020b), ensemble83.334.60
Koo et al. (2020), VGGish72.925.07
Koo et al. (2020), Transformer-XL81.254.01
Koo et al. (2020), VGGish + GloVe77.084.33
Koo et al. (2020), VGGish + transformer-XL75.003.74
Koo et al. (2020), ensembled output81.253.77
Campbell et al. (2020), fusion II75.00
Campbell et al. (2020), fusion I72.92
Campbell et al. (2020), RNN model75.00
Campbell et al. (2020), fluency60.42
Campbell et al. (2020), x-vector54.17
Sarawgi et al. (2020a), UA ensemble4.35
Sarawgi et al. (2020a), UA ensemble (weighted)3.93
Pappagari et al. (2020), acoustic and transcript75.005.37
Rohanian et al. (2020), LSTM (Lexical + Dis)72.924.88
Rohanian et al. (2020), LSTM with gating (Acoustic + Lexical)77.084.57
Rohanian et al. (2020), LSTM with gating (Acoustic + Lexical + Dis)79.174.54
Yuan et al. (2020), ERNIE3p89.58
Syed et al. (2020)85.424.30
Edwards et al. (2020), phonemes and audio79.17
Meghanani et al. (2021), CNN-LSTM with MFCC64..586.24
Meghanani et al. (2021), pBLSTM-CNN with log-Mel52.085.90
Meghanani et al. (2021), ResNet-LSTM with log-Mel62.505.98

Baseline methods on ADReSS test set.

3.1 Acoustic Feature-Based Methods

Luz et al. (2020) explore several acoustic features like extended Geneva minimalistic acoustic parameter set (eGeMAPS) (Eyben et al., 2016), emobase, ComParE-2013 (Eyben et al., 2013), and multiresolution cochleagram (MRCG) (Chen et al., 2014), feeding the traditional machine learning algorithms like linear discriminant analysis, decision trees, nearest neighbor, random forests, and support vector machines. In our previous work (Meghanani et al., 2021), we have used CNN/ResNet + long short-term memory (LSTM) networks and pyramidal bidirectional LSTM + CNN networks trained on log-Mel spectrogram and Mel-frequency cepstral coefficient (MFCC) features extracted from the spontaneous speech. Pompili et al. (2020) exploit the pretrained models to produce i-vector- and x-vector-based acoustic feature embeddings. They evaluate x-vector, i-vector, and statistical speech-based functional features. Rhythmic features are proposed in Campbell et al. (2020), as lower speaking fluency is a common pattern in patients with AD. Koo et al. (2020) use VGGish (Hershey et al., 2017) trained with Audio Set (Gemmeke et al., 2017) for audio classification. They have proposed a modified version of convolutional recurrent neural network (CRNN), where an attention layer is the forefront layer of the network, and fully connected layers follow the recurrent layer.

3.2 Linguistic Feature-Based Methods

Recently, there have been multiple attempts on the AD detection problem based on text-based features and models. Searle et al. (2020) use traditional machine learning techniques like support vector machines (SVMs), gradient boosting decision trees (GBDT), and conditional random fields (CRFs). They also try deep learning transformer-based models, specifically, bidirectional encoder representations from transformers (BERT) (Devlin et al., 2019), RoBERTa (Liu et al., 2019), and DistilBERT/DistilRoBERTa (Sanh et al., 2019). Pompili et al. (2020) encode each word of the clean transcriptions into 768-dimensional context embedding vector using a frozen English BERT model pretrained with 12 layers. Three different neural models are trained on top of contextual word embeddings: 1) global maximum pooling, 2) bidirectional long short-term memory (BLSTM)–based recurrent neural networks (RNN) provided with an attention module, and 3) the second model augmented with part-of-speech (POS) embeddings. In the work of Campbell et al. (2020), authors have used the manual transcripts to extract linguistic information (interventions, vocabulary richness, frequency of verbs, nouns, POS-tagging, etc.) for creating the input features of the classifier. They use another sequential deep learning-based classifier, which directly classifies the sequence of Gobal Vectors (GloVe)–based word embeddings. Koo et al. (2020) use transformer-based language models (Vaswani et al., 2017), generative pretraining (GPT) (Radford et al., 2018), RoBERTa (Liu et al., 2019), and transformer-XL (Dai et al., 2020) to get textual features and perform classification and regression tasks using a modified convolutional recurrent neural network-based structure.

Graph-based representation of word features (Tomás and Radev, 2012; Cong and Liu, 2014), which have shown promise in classifying texts (De Arruda et al., 2016), is also employed for detection of mild cognitive impairments. Santos et al. (2017) model transcripts as complex networks and enrich them with word embedding to better represent short texts produced in neuropsychological assessments. They use metrics of topological properties of complex networks in a machine learning classification approach to distinguish between healthy subjects and patients with mild cognitive impairments. Such graph-based techniques have also been used in the word sense disambiguation (WSD) tasks to identify the meaning of words in a given context for specific words conveying multiple meanings. Corra et al. (2018) suggest that a bipartite network model with local features employed to characterize the context can be useful in improving the semantic characterization of written texts without the use of deep linguistic information.

3.3 Bimodal Methods

Methods with bimodal input features (both acoustic and linguistic) are also used for AD recognition in various studies (Sarawgi et al., 2020a; Sarawgi et al., 2020b; Campbell et al., 2020; Koo et al., 2020; Pompili et al., 2020; Rohanian et al., 2020). However, in this work, we restrict ourselves to the NLP-based approaches.

4 Proposed NLP-Based Methods

4.1 Data Preparation

In this work, we explore the linguistic features for AD detection and hence only the textual transcripts in the ADReSS dataset are used. The transcripts contain the conversational content between the participant and the investigator. This includes pauses in speech, laughter, and discourse markers such as “um” and “uh.” Each transcript is considered as a single data point with their corresponding AD label and MMSE score. We create two transcription level datasets after preprocessing the transcripts as in Searle et al. (2020)—1) PAR: containing the utterances of participant alone, 2) PAR + INV: containing utterances from both the participant and the investigator. In addition to the preprocessing performed in Searle et al. (2020), we keep PAR and INV tags as well in the data (which defines whether the utterance is spoken by the participant or the investigator).

4.2 Convolutional Neural Network Model

Language impairments like difficulties in lexical retrieval, loss of verbal fluency, and breakdown in comprehension of higher order written and spoken languages are common in AD patients. Hence the linguistic information, like the n-grams present in the input sentence, may provide good cues for AD detection. Any CNN filter, where n is the number of sequential words looked over by the filter and d is the dimension of word embedding, can be viewed as a feature detector looking for a specific n-gram in the input that can capture the language impairments associated with AD.

We describe the details of the CNN model from the work (Kim, 2014) as follows. Let be a d-dimensional word vector corresponding to the ith word in the sentence. A sentence of length L is represented as . Let represent the concatenation of the words . A convolution operation involves a filter , which is applied to a window of n words to produce a new feature as shown in Eq. 3, where is generated from a window of words by

In Eq. 3, f is a nonlinear function and b is the bias term. A feature map is obtained by applying the filter to all possible windows of words in the sentence .

A max-pool over time (Collober et al., 2011) is performed over the feature map to get as the feature corresponding to that filter. This corresponds to the n-gram that is “most relevant” in the AD recognition task. The weights of the filters, which in turn determine the “most relevant” feature, are learnt using backpropagation. CNNs are trained with just one layer of convolution. Variable length sentences are automatically handled by the pooling scheme. We use pretrained 100-dimensional GloVe word vectors (Pennington et al., 2014) for word embedding. Multiple kernels of sizes , , , and are employed to have a look at the bigrams, trigrams, 4-grams, and 5-grams within the text. We use 100 filters each with heights 2, 3, 4, and 5. Multiple configurations with filter sizes [2,3,4], [3,4,5], and [2,3,4,5] are applied which are referred to as CNN-bi+tri+4 gram, CNN-tri+4+5 gram, and CNN-bi+tri+4+5 gram in our tables. The outputs of the filter are concatenated together to form a single vector. Dropout with probability is applied on the concatenated filter output and the results are passed through a linear layer for the final prediction task. The linear layer weights up the evidence from each of these n-grams and make a final decision. Figure 1 shows the basic CNN operation over an example sentence.

FIGURE 1

4.2.1 Training Details

For the classification task, training is performed for 100 epochs with a batch size of 16. Adam optimizer is used with a learning rate of 0.001. Model with the lowest validation loss is saved and used for prediction. Since AD classification is a two-class problem, binary cross-entropy with logits loss is used as the loss function. For the MMSE score prediction task, the output layer is a fully connected layer with linear activation function. In the regression task the network is trained for 1,500 epochs with the objective to minimize the mean squared error.

We use bootstrap aggregation of models known as bagging (Breiman, 1996) to predict the final labels/MMSE scores for test samples. Bootstrap aggregation is an ensemble technique to improve the stability and accuracy of machine learning models. It combines the prediction from multiple models. It also reduces variance and helps to avoid overfitting. We fit 21 models and the outputs are combined by a majority voting scheme for final classification. In the regression task, the outputs of these bootstrap models are averaged to arrive at the final MMSE score.

4.3 fastText

fastText-based classifiers calculate the n-grams of an input sentence explicitly and append them to the end of the sentence. In this work, we use bigrams and trigrams. We conducted the experiments with 4-grams as well, but the results did not show any improvement over the use of trigrams. This bag of bigrams and trigrams acts as additional features to capture some information about the local word order.

Figure 2 shows the architecture of fastText model. The fastText model has two layers, an embedding layer and a linear layer. The embedding layer calculates the word embedding (100-dimensional) for each word. The average of all these word embeddings is calculated and fed through the linear layer for final prediction as described in Figure 2. fastText models are faster for training and evaluation by many orders of magnitude, compared to the “deep” models. As mentioned in the work (Joulin et al., 2017), fastText can be trained on more than one billion words in less than 10 min using a standard multicore CPU and classify half a million sentences among 312 K classes in less than a minute.

FIGURE 2

4.3.1 Training Details

All training details are the same as mentioned in Section 4.2.1. The only difference is that dropout is not used in this model. Here also we use 21 bootstrapping models and the outputs are combined as described in Section 4.2.1.

5 Results

We have performed 5-fold cross-validation, to estimate the generalization error. One of the folds has 20 validation samples and the remaining four have 22 validation samples. The results of cross-validation on CNN and fastText models trained on PAR and PAR + INV sets are listed in Table 2. The best performing model for classification during the cross-validation was fastText with bigrams on the PAR + INV set, which yields an average cross-validation accuracy of 86.09%. Among the CNN models, tri+4+5 grams give the best accuracy in both PAR (77.54%) and INV + PAR (81.27%) sets. As far as accuracy is concerned, both the CNN and fastText models seem to benefit from the inclusion of utterances from the investigator. For the prediction of MMSE score, CNN with bi+tri+4+5 grams (RMSE of 4.38) was the best. The fastText models seem to get a clear advantage in RMSE with the addition of the utterances from the investigator. However such a large difference in RMSE is not observable between the CNN models using PAR and INV + PAR sets. The cross-validation results confirmed our belief that the n-grams from the transcriptions of the picture description task could be useful in the detection of AD.

TABLE 2

DatasetModelAccuracyRMSE
PARCNN, bi+tri+4 gram73.914.55
PARCNN, tri+4+5 gram77.544.41
PARCNN, bi+tri+4+5 gram76.544.65
PARfastText, bigram80.545.43
PARfastText, bi + trigram82.365.40
PAR + INVCNN, bi+tri+4 gram80.184.63
PAR + INVCNN, tri+4+5 gram81.274.53
PAR + INVCNN, bi+tri+4+5 gram80.364.38
PAR + INVfastText, bigram86.094.66
PAR + INVfastText, bi + trigram85.904.81

Average 5-fold cross-validation results for AD classification and RMSE values.

Table 3 lists the classification accuracy and RMSE in the prediction of MMSE score on the test set of the ADReSS corpus. The table also lists the precision, recall, and for each class. They are computed as precision , recall , and , where , , , and are the number of true positives, false positives, true negatives, and false negatives, respectively. The listed results are obtained after bootstrapping with 21 samples. The best classification accuracy is 83.33% which is achieved using fastText model with appended bigrams and trigrams. The accuracies are similar in both PAR and PAR + INV sets using the fastText model. The maximum accuracy obtained with CNN models is 79.16%, which is achieved on the INV + PAR set using bi+tri+4 grams or tri+4+5 grams. In the detection task, the CNN models seem to benefit from the addition of utterances from the investigator. Also the accuracies seem to degrade when bigrams, trigrams, 4-grams, and 5-grams are considered together. This behavior is consistent across the PAR and PAR + INV sets. The best RMSE in the prediction of MMSE score is 4.28 which is obtained on the PAR + INV set using fastText model employing only bigrams. In the regression task using fastText, the use of bigrams achieves slightly better RMSE compared to the use of both bigrams and trigrams. Also the fastText models seem to benefit from the use of utterances from the investigator. In contrast, CNN models do not seem to get any specific advantage with the inclusion of investigator’s utterances. The performance of the CNN models remains almost the same across the use of bi+tri+4, tri+4+5, and bi+tri+4+5 grams.

TABLE 3

DatasetModelClassPrecisionRecallF1 scoreAccuracy (%)RMSE
PARCNN, bi+tri+4 gramNon-AD0.740.710.7272.914.38
AD0.720.750.73
PARCNN, tri+4+5 gramNon-AD0.760.670.7172.914.46
AD0.700.790.75
PARCNN, bi+tri+4+5 gramNon-AD0.710.710.7170.834.42
AD0.710.710.71
PARfastText, bigramNon-AD0.780.880.8281.254.51
AD0.860.750.80
PARfastText, bi + trigramNon-AD0.810.880.8483.334.87
AD0.860.790.83
PAR + INVCNN, bi+tri+4 gramNon-AD0.770.830.8079.164.48
AD0.820.750.78
PAR + INVCNN, tri+4+5 gramNon-AD0.770.830.8079.164.47
AD0.820.750.78
PAR + INVCNN, bi+tri+4+5 gramNon-AD0.740.710.7272.914.44
AD0.720.750.73
PAR + INVfastText, bigramNon-AD0.780.880.8281.254.28
AD0.860.750.80
PAR + INVfastText, bi + trigramNon-AD0.790.920.8583.334.47
AD0.900.750.82

Results on ADReSS test set. The bold values represent the best results obtained by our models.

6 Discussion and Conclusion

In this work, we explore two models, CNN with a single convolution layer and fastText, to address the problem of AD classification and prediction of MMSE score from the transcriptions of the picture description task. The choice of these models was based on our initial belief that modeling the transcriptions of the narrative speech in the picture description task using n-grams could give some indication on the status of AD. The chosen models are also shallow. The number of parameters is much less than the usual deep learning architectures and hence they can be trained and evaluated quite fast. Yet, the performance of these models is competitive with the baseline results reported with complex models (refer to Table 1). The results suggest that the n-gram-based features are worth pursuing, for the task of AD detection.

Among the considered models, fastText model with bigrams and trigrams appended to the input achieves the best classification accuracy (83.33%). In the regression task, the best results (RMSE of 4.28) are achieved using fastText model with only the bigrams appended to the input. The fastText models have a clear edge over CNN in the classification task. Empirical evidence suggests that fastText models benefit from the inclusion of utterances from the investigator in the regression task, though they do not make much difference in the classification task. The CNN models on the other hand perform better on the PAR + INV sets in the classification task. In the regression task, their performance is similar across the PAR and PAR + INV sets. Bigrams have an edge over bi + tri grams in fastText, when used for prediction of MMSE score. However, the performance of the CNN models remains almost the same across the use of bi+tri+4, tri+4+5, and bi+tri+4+5 grams, in the regression task.

Statements

Data availability statement

The data analyzed in this study are subject to the following licenses/restrictions: In order to gain access to the ADReSS data, you will need to become a member of DementiaBank (free of charge) by contacting Brian MacWhinney on . You should include your contact information and affiliation, as well as a general statement on how you plan to use the data, with specific mention to the ADReSS challenge. If you are a student, please ask your supervisor to join as a member as well. This membership will give you full access to the DementiaBank database, where the ADReSS dataset will be available and clearly identified. For further information, visit DementiaBank. Requests to access these datasets should be directed to Brian MacWhinney, .

Author contributions

AM, AS, and AR contributed to the conception and design of the study. AM and AS wrote the first draft of the manuscript. AR reviewed the first draft and suggested improvements. AM and AS wrote sections of the manuscript. All authors contributed to manuscript revision and read and approved the submitted version.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  • 1

    BreimanL. (1996). Bagging predictors. Mach. Learn.24, 123140. 10.1007/BF00058655

  • 2

    CampbellE. L.Docío-FernándezL.RabosoJ. J.García-MateoC. (2020). Alzheimer’s dementia detection from audio and text modalities. arXiv preprint arXiv:2008.04617

  • 3

    ChenJ.WangY.WangD. (2014). A feature study for classification-based speech separation at low signal-to-noise ratios. IEEE/ACM Trans. Audio Speech Lang. Process.22, 19932002. 10.1109/TASLP.2014.2359159

  • 4

    ColloberR.WestonJ.BottouL.KarlenM.KavukcuogluK.KuksaP. (2011). Natural language processing (almost) from scratch. J. Machine Learn. Res.12, 24932537. 10.5555/1953048.2078186

  • 5

    CongJ.LiuH. (2014). Approaching human language with complex networks. Phys. Life Rev.11, 598618. 10.1016/j.plrev.2014.04.004

  • 6

    CorraE. A.LopesA. A.AmancioD. R. (2018). Word sense disambiguation. Inf. Sci.442, 103113. 10.1016/j.ins.2018.02.047

  • 7

    DaiZ.YangZ.YangY.CarbonellJ.LeQ.SalakhutdinovR. (2019). “Transformer-XL: attentive language models beyond a fixed-length context,” in Proceedings of the 57th annual meeting of the association for computational linguistics, Florence, Italy, July 2019, 29782988. 10.18653/v1/P19-1285

  • 8

    De ArrudaH.CostaL.AmancioD. (2016). Using complex networks for text classification: discriminating informative and imaginative documents. EPL113, 28007. 10.1209/0295-5075/113/28007

  • 9

    DevlinJ.ChangM.-W.LeeK.ToutanovaK. (2019). “BERT: pre-training of deep bidirectional transformers for language understanding,” in Proceedings of the 2019 conference of the north american chapter of the association for computational linguistics: human language technologies, Minneapolis, MN, June 2–7, 2019, Vol. 1, 41714186. 10.18653/v1/N19-1423

  • 10

    EdwardsE.DogninC.BollepalliB.SinghM. (2020). “Multiscale system for Alzheimer’s dementia recognition through spontaneous speech,” in Proceedings of interspeech 2020, Shanghai, China, October 2020, 21972201. 10.21437/Interspeech.2020-2781

  • 11

    EybenF.SchererK. R.SchullerB. W.SundbergJ.AndréE.BussoC.et al (2016). The Geneva minimalistic acoustic parameter set (GeMAPS) for voice research and affective computing. IEEE Trans. Affective Comput.7, 190202. 10.1109/taffc.2015.2457417

  • 12

    EybenF.WeningerF.GrossF.SchullerB. (2013). “Recent developments in openSMILE, the Munich open-source multimedia feature extractor,” in Proceedings the 2013 ACM multimedia conference, Barcelona, Spain, October, 2013, 835838. 10.1145/2502081.2502224

  • 13

    GemmekeJ.EllisD.FreedmanD.JansenA.LawrenceW.MooreR.et al (2017). “Audio set: an ontology and human-labeled dataset for audio events,” in 2017 IEEE international conference on acoustics, speech and signal processing (ICASSP), New Orleans, LA, March 5–9, 2017, 776780. 10.1109/ICASSP.2017.7952261

  • 14

    HersheyS.ChaudhuriS.EllisD.GemmekeJ.JansenA.MooreR. C.et al (2017). “CNN architectures for large-scale audio classification,” in 2017 IEEE international conference on acoustics, speech and signal processing (ICASSP), New Orleans, LA, March 5–9, 2017, 131–(135.)

  • 15

    JoulinA.GraveE.BojanowskiP.MikolovT. (2017). “Bag of tricks for efficient text classification,” in Proceedings of the 15th conference of the european chapter of the association for computational linguistics, Valencia, Spain, April 3–7, 2017, Vol. 2, 427–(431.)

  • 16

    KimY. (2014). “Convolutional neural networks for sentence classification,” in Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), Doha, Qatar, October 25–29, 2014, 17461751. 10.3115/v1/D14-1181

  • 17

    KooJ.LeeJ. H.PyoJ.JoY.LeeK. (2020). “Exploiting multi-modal features from pre-trained networks for Alzheimer’s dementia recognition,” in Proceedings of interspeech 2020, Shanghai, China, October 2020, 22172221. 10.21437/Interspeech.2020-3153

  • 18

    LiuY.OttM.GoyalN.DuJ.JoshiM.ChenD.et al (2019). RoBERTa: a robustly optimized BERT pretraining approach. ArXiv abs/1907.11692

  • 19

    LuzS.HaiderF.de la FuenteS.FrommD.MacWhinneyB. (2020). “Alzheimer’s dementia recognition through spontaneous speech: the ADReSS challenge,” in Proceedings of interspeech 2020, Shanghai, China, October 2020, 21722176. 10.21437/Interspeech.2020-2571

  • 20

    MacwhinneyB. (2009). “The CHILDES project part 1,” in The CHAT transcription format. 10.1184/R1/6618440.v1

  • 21

    MeghananiA.AnoopC. S.RamakrishnanA. G. (2021). “An exploration of log-mel spectrogram and MFCC features for Alzheimer’s dementia recognition from spontaneous speech,” in The 8th IEEE spoken language technology workshop (SLT), Shenzhen, China, January 19-22, 2021

  • 22

    MuellerK. D.HermannB.MecollaribJ.TurkstraL. S. (2018a). Connected speech and language in mild cognitive impairment and Alzheimer’s disease: a review of picture description tasks. J. Clin. Exp. Neuropsychol.40, 917939. 10.1080/13803395.2018.1446513

  • 23

    MuellerK. D.KoscikR. L.HermannB.JohnsonS. C.TurkstraL. S. (2018b). Declines in connected language are associated with very early mild cognitive impairment: results from the Wisconsin registry for Alzheimer’s prevention. Front. Aging Neurosci.9, 437. 10.3389/fnagi.2017.00437

  • 24

    NicholasM.OblerL. K.AlbertM.Helm-EstabrooksN. (1985). Empty speech in Alzheimer’s disease and fluent aphasia. J. Speech Hear. Res.28, 405410. 10.1044/jshr.2803.405

  • 25

    PappagariR.ChoJ.Moro-VelázquezL.DehakN. (2020). “Using state of the art speaker recognition and natural language processing technologies to detect Alzheimer’s disease and assess its severity,” in Proceedings of interspeech 2020, Shanghai, China, October 2020, 21772181. 10.21437/Interspeech.2020-2587

  • 26

    PenningtonJ.SocherR.ManningC. (2014). Glove: global vectors for word representation. Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), Doha, Qatar, October 25–29, 2014, 15321543. 10.3115/v1/d14-1162

  • 27

    PompiliA.RollandT.AbadA. (2020). “The INESC-ID multi-modal system for the ADReSS 2020 challenge,” in Proceedings of interspeech 2020, Shanghai, China, October 2020, 22022206. 10.21437/Interspeech.2020-2833

  • 28

    RadfordA.NarasimhanK.SalimansT.SutskeverI. (2018). Improving language understanding by generative pre-training. Available at: https://www.cs.ubc.ca/amuham01/LING530/papers/radford2018improving.pdf. 10.1017/9781108552202

  • 29

    RohanianM.HoughJ.PurverM. (2020). “Multi-Modal fusion with gating using audio, lexical and disfluency features for Alzheimer’s dementia recognition from spontaneous speech,” in Proceedings of interspeech 2020, Shanghai, China, October 2020, 21872191. 10.21437/Interspeech.2020-2721

  • 30

    SanhV.DebutL.ChaumondJ.WolfT. (2019). DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. ArXiv abs/1910.01108

  • 31

    SantosL.Corrêa JúniorE. A.OliveiraO.Jr.AmancioD.MansurL.AluísioS. (2017). “Enriching complex networks with word embeddings for detecting mild cognitive impairment from speech transcripts,” in Proceedings the 55th annual meet: the association for computational linguistics, Vancouver, BC, July 30–August 4, 2017, Vol. 1, 12841296. 10.18653/v1/P17-1118

  • 32

    SarawgiU.ZulfikarW.KhinchaR.MaesP. (2020a). Uncertainty-aware multi-modal ensembling for severity prediction of Alzheimer’s dementia. ArXiv abs/2010.01440. 10.21437/interspeech.2020-3137

  • 33

    SarawgiU.ZulfikarW.SolimanN.MaesP. (2020b). Multimodal inductive transfer learning for detection of Alzheimer’s dementia and its severity. arXiv preprint arXiv:2009.00700. 10.21437/interspeech.2020-3137

  • 34

    SavundranayagamM.HummertM. L.MontgomeryR. (2005). Investigating the effects of communication problems on caregiver burden. J. Gerontol. B Psychol. Sci. Soc. Sci.60 (1), S48S55. 10.1093/geronb/60.1.s48

  • 35

    SearleT.IbrahimZ.DobsonR. (2020). “Comparing natural language processing techniques for Alzheimer’s dementia prediction in spontaneous speech,” in Proceedings of interspeech 2020, Shanghai, China, October 2020, 21922196. 10.21437/Interspeech.2020-2729

  • 36

    ShenY.HeX.GaoJ.DengL.MesnilG. (2014). “Learning semantic representations using convolutional neural networks for web search,” in WWW 2014, Seoul, South Korea, April 7–11, 2014, 373374. 10.1145/2567948.2577348

  • 37

    SyedM. S. S.SyedZ. S.LechM.PirogovaE. (2020). “Automated screening for Alzheimer’s dementia through spontaneous speech,” Proceedings of interspeech 2020, Shanghai, China, October 2020, 22222226. 10.21437/Interspeech.2020-3158

  • 38

    SzatlóczkiG.HoffmannI.VinczeV.KálmánJ.PákáskiM. (2015). Speaking in Alzheimer’s disease, is that an early sign? Importance of changes in language abilities in Alzheimer’s disease. Front. Aging Neurosci.7, 110. 10.3389/fnagi.2015.00195

  • 39

    tau YihW.HeX.MeekC. (2014). “Semantic parsing for single-relation question answering,” in Proceedings of the 52nd annual meeting of the association for computational linguistics, Baltimore, MA, June 2014, Vol. 2, 643648. 10.3115/v1/P14-2105

  • 40

    TomásD. R. M.RadevD. (2012). Graph-based natural language processing and information retrieval. Machine Translation26, 277280. 10.1007/s10590-011-9122-9

  • 41

    VaswaniA.ShazeerN.ParmarN.UszkoreitJ.JonesL.GomezA.et al (2017). “Attention is all you need,” Proceedings of the 31st international conference on neural information processing systems, Long Beach, CA, December 2017, 5999–(6009.)

  • 42

    YuanJ.BianY.CaiX.HuangJ.YeZ.ChurchK. (2020). “Disfluencies and fine-tuning pre-trained language models for detection of Alzheimer’s disease,” in Proceedings of interspeech 2020, Shanghai, China, October 2020, 21622166. 10.21437/Interspeech.2020-2516

Summary

Keywords

fastText, convolutional neural network, Alzheimer’s, dementia, mini-mental state examination

Citation

Meghanani A, Anoop CS and Ramakrishnan AG (2021) Recognition of Alzheimer’s Dementia From the Transcriptions of Spontaneous Speech Using fastText and CNN Models. Front. Comput. Sci. 3:624558. doi: 10.3389/fcomp.2021.624558

Received

31 October 2020

Accepted

25 January 2021

Published

24 March 2021

Volume

3 - 2021

Edited by

Saturnino Luz, University of Edinburgh, United Kingdom

Reviewed by

Diego R. Amancio, University of São Paulo, Brazil

Anna Pribilova, Slovak Academy of Sciences (SAS), Slovakia

Updates

Copyright

*Correspondence: C. S. Anoop,

This article was submitted to Human-Media Interaction, a section of the journal Frontiers in Computer Science

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics