Skip to main content

ORIGINAL RESEARCH article

Front. Psychiatry, 28 April 2022
Sec. Mood Disorders
This article is part of the Research Topic The Development of Biomarkers in Psychiatry View all 11 articles

Vocal Acoustic Features as Potential Biomarkers for Identifying/Diagnosing Depression: A Cross-Sectional Study

\r\nQing ZhaoQing ZhaoHong-Zhen FanHong-Zhen FanYan-Li LiYan-Li LiLei LiuLei LiuYa-Xue WuYa-Xue WuYan-Li ZhaoYan-Li ZhaoZhan-Xiao TianZhan-Xiao TianZhi-Ren WangZhi-Ren WangYun-Long TanYun-Long TanShu-Ping Tan*\r\nShu-Ping Tan*
  • Peking University HuiLongGuan Clinical Medical School, Beijing Huilongguan Hospital, Beijing, China

Background: At present, there is no established biomarker for the diagnosis of depression. Meanwhile, studies show that acoustic features convey emotional information. Therefore, this study explored differences in acoustic characteristics between depressed patients and healthy individuals to investigate whether these characteristics can identify depression.

Methods: Participants included 71 patients diagnosed with depression from a regional hospital in Beijing, China, and 62 normal controls from within the greater community. We assessed the clinical symptoms of depression of all participants using the Hamilton Depression Scale (HAMD), Hamilton Anxiety Scale (HAMA), and Patient Health Questionnaire (PHQ-9), and recorded the voice of each participant as they read positive, neutral, and negative texts. OpenSMILE was used to analyze their voice acoustics and extract acoustic characteristics from the recordings.

Results: There were significant differences between the depression and control groups in all acoustic characteristics (p < 0.05). Several mel-frequency cepstral coefficients (MFCCs), including MFCC2, MFCC3, MFCC8, and MFCC9, differed significantly between different emotion tasks; MFCC4 and MFCC7 correlated positively with PHQ-9 scores, and correlations were stable in all emotion tasks. The zero-crossing rate in positive emotion correlated positively with HAMA total score and HAMA somatic anxiety score (r = 0.31, r = 0.34, respectively), and MFCC9 of neutral emotion correlated negatively with HAMD anxiety/somatization scores (r = −0.34). Linear regression showed that the MFCC7-negative was predictive on the PHQ-9 score (β = 0.90, p = 0.01) and MFCC9-neutral was predictive on HAMD anxiety/somatization score (β = −0.45, p = 0.049). Logistic regression showed a superior discriminant effect, with a discrimination accuracy of 89.66%.

Conclusion: The acoustic expression of emotion among patients with depression differs from that of normal controls. Some acoustic characteristics are related to the severity of depressive symptoms and may be objective biomarkers of depression. A systematic method of assessing vocal acoustic characteristics could provide an accurate and discreet means of screening for depression; this method may be used instead of—or in conjunction with—traditional screening methods, as it is not subject to the limitations associated with self-reported assessments wherein subjects may be inclined to provide socially acceptable responses rather than being truthful.

Introduction

Depression is a serious mental illness that can result in a significant decline in a patient’s social functioning and quality of life. According to a 2019 epidemiological survey in China, the lifetime prevalence of depression was 3.4%, and the 12-month prevalence was 2.1%—both of which are higher than the prevalence rates for most other psychiatric disorders (1). Currently, the diagnosis of depression is primarily dependent upon the subjective judgment of experienced psychiatrists. However, accurately distinguishing negative emotions from depression is challenging, and objective and effective auxiliary diagnostic methods for depression are lacking. These are among the possible reasons that depression is easily missed. For example, in primary care settings, only one-third of people with mild depression are correctly diagnosed (2). After experiencing a first depressive episode, the recurrence rate within 3 years is more than 30%, and some patients have recurrent episodes; thus, the course of the disease can be chronic, placing a heavy burden on both patients and society (3, 4). Consequently, the use of any potential objective biomarker for accurately diagnosing depression warrants in-depth investigation.

The speech of depressed patients is different from that of normal people. The patients usually show lower intonation and less speech. Many studies have suggested that there are abnormal acoustic features in depressed patients (5, 6). This suggests that acoustic features may be an objective biomarker of depression. Therefore, to determine an appropriate biomarker to investigate, we considered the question, “What is the physical process of speech?” The brain organizes prosodic information and produces neuromuscular instructions that control the activities of muscles and tissues related to phonation movement. Next, the airflow stream out of the lungs either causes the vocal cords to vibrate (when the glottis is closed) or passes through the vocal cord smoothly (when the glottis is open). The oropharyngeal muscle forms the main channel of phonation, which is equivalent to a filter that can amplify or attenuate the sound of a specific frequency (7).

Regarding the content of speech, it mainly consists of two different types of information: “what” is being said, including the literal meaning of the information expressed, and “how” it is said, which connotes the emotion and rhythm information in the expression. The speech of a depressed patient can convey not only verbal information, such as negative, pessimistic, and suicidal ideas, but also considerable non-verbal information, including slow reaction and speech, lack of attention, and low volume of voice. “Voice,” in this context, is a non-verbal form of emotional expression (8). Emotion affects speech by changing the physiological characteristics of the vocal system (9), breathing, and muscle tone while listeners perceive the emotion of the speaker through specific neural pathways (10).

Depressed patients tend to show less speech production, have a low intonation, and pause often, all of which are characteristics of this mental disorder. Acoustic characteristics of speech that people can hear directly, such as the length of a pause, pause variability, and the speech/pause ratio, differ between people with and without depression, and studies have verified that some of these characteristics indicate depression severity (1114). These reported findings suggest that voice analysis may be used to gauge the severity of depression objectively. However, the aforementioned characteristics are easily affected by unstable environmental factors. Therefore, the mel-frequency cepstral coefficient (MFCC) is often considered a superior method for distinguishing differences in voice emotion characteristics and analyzing the subtle differences in voice emotions perceived by human ears (15).

MFCCs can be used to identify changes in the vocal tract (16) and can be calculated in the following steps: (1) calculating the fast Fourier transform spectrum from the frequency, (2) extracting filter bank output of mel scale allocation according to human auditory features, and (3) obtaining the cepstrum coefficient from the discrete cosine transformation. Ozdas et al. (17) found that MFCC has an accuracy of 75% in distinguishing between depressed patients and normal controls. Akkaralaertsest and Yingthawornsuk (18) also believed that there were characteristic differences in MFCC of depressed patients’ speech that could help identify depressed patients with or without high suicide risk, to some extent.

In addition to the MFCC, other characteristics can indicate the voice mood. The fundamental frequency (F0) is the basic frequency generated by the vibration of the vocal cords. Results from a 4-week sertraline/placebo double-blind controlled trial showed that the mean fundamental frequency of the depression group that responded to the treatment increased from baseline; meanwhile, the frequency remained consistent or decreased from baseline for the depression group that did not respond to the treatment (19). Additionally, the zero-crossing rate (ZCR), which refers to the number of times a signal passes through the zero axis in a certain time, can be used to identify voiceless and voiced sounds that are characteristic expressions of phonetic emotions. A Chinese study found the ZCR has an accuracy of 67–70% to distinguish depressed individuals from those without depression (20). However, a Japanese study found no significant difference in ZCR between the depression and normal groups (5). Furthermore, although the harmonic-to-noise ratio (HNR) reflects the ratio of the sound signal strength to noise signal strength, few studies have investigated the relationship between depressive symptoms and acoustic characteristics, such as ZCR, MFCC, HNR, and F0.

Additionally, although previous studies that used small sample sizes and included limited acoustic characteristics have shown that depressed patients present abnormal acoustic characteristics, no studies to date have verified the correlation between acoustic characteristics and clinical symptoms. Hence, there is a lack of explanation for the psychopathology of abnormal acoustic characteristics. It is unclear whether acoustic features can partly reflect the severity of emotional symptoms and whether the severity of depression can be evaluated by certain acoustic features.

To address this limitation, the present study surveyed a relatively large sample of depressed patients and healthy controls. Using MFCC, ZCR, HNR, and F0, which are known to reflect the emotional changes of the voice, we aimed to verify the relationship between acoustic characteristics and severity of depression and the significance of acoustic characteristics in disease discrimination. We also conducted a preliminary exploration of the psychopathological mechanism of abnormal acoustic characteristics.

We hypothesized that the acoustic characteristics of depressed patients and normal people differ and are related to depressive symptoms and that acoustic characteristics can help distinguish depressed patients from normal individuals to some extent. If acoustic characteristics can be used as objective biomarkers of depression to provide a diagnostic reference, it may help decrease the rate of misdiagnosis and missed diagnosis, promote early identification and treatment of depression, and ultimately improve the patient’s prognosis. Since acoustic characteristics can be obtained automatically, exploring the acoustic characteristics that can predict the severity of anxiety and depression was conducive to promoting the intelligent evaluation of depressive disorder severity. Moreover, it can also promote the development of artificial intelligence ability to identify depression and provide Internet-based medical treatment to patients, which can have far-reaching implications.

Materials and Methods

Participants

We recruited 71 patients from Beijing Huilongguan Hospital from 2018 to 2020. An experienced psychiatrist conducted detailed mental and physical examinations of the participants. The inclusion criteria were as follows: (1) diagnosed with depression according to the Diagnostic and Statistical Manual of Mental Disorders (DSM–IV), (2) aged 18–60 years, (3) attained junior high school education or above, and (4) able to read Chinese fluently. The exclusion criteria were as follows: (1) diagnosed with severe cerebrovascular disease, brain injury, and other nervous system diseases; (2) having a serious physical disease; (3) having accompanying substance abuse (except tobacco) problems; and (4) diagnosed with acute nasolaryngology and respiratory system diseases affecting the vocal cords. Sixty-two healthy controls with no current or previous mental illness were recruited from the surrounding community and screened by the Mini-International Neuropsychiatric Interview. The remaining inclusion and exclusion criteria were the same as those used in the patient group.

This study was approved by the Ethics Committee of Beijing Huilongguan Hospital. A specific psychiatrist explained the details of the study to all the participants; meanwhile, their information was strictly confidential. Audio data would be stored on a specific hard disk. Participants took part in voluntarily and there would be no penalty for withdrawal at any time. Written informed consent was obtained from all participants.

Measures

All participants provided their general information, and the depression group completed clinical symptom evaluation questionnaires, including the 17-item Hamilton Depression Scale (HAMD-17), Hamilton Anxiety Scale (HAMA), and Patient Health Questionnaire (PHQ-9). The HAMD-17 is a commonly used depression rating scale comprising 17 symptom items across five domains including anxiety/somatization, weight, cognitive impairment, retardation, sleep disturbance (21, 22). The HAMA is a commonly used 14-item anxiety rating scale for measuring “physical anxiety” and “mental anxiety;” each item is scored from 0 (not present) to 4 (very severe) (23). We also used the PHQ-9, a simple and effective 9-item self-assessment tool for depression wherein each item is rated on a scale from 0 to 3 (24). Higher scores on these scales indicate greater severity of symptoms.

All participants were seated in a noise-controlled room (background sound less than 30 dB). After a 3-min rest period, voices were recorded using a fixed ISK BM-5000 microphone (sampling frequency: 44 kHz, bit rate: 1,058 kbps). The participants were asked to read three paragraphs in their natural tone. Each paragraph contained about 200 Chinese characters, which described one of the three different emotions: positive, negative, and neutral. The positive emotion text depicted scenes from the Chinese New Year and included words such as “happy,” “smile,” “play games,” “delicious,” “singing and dancing,” and so on. The negative emotion text included “punishment,” “trouble,” “worry,” “anxious,” “worry,” “slander,” “pain,” “dark clouds,” and so on. The neutral emotion text described a bridge using location, shape, and other descriptors that did not contain any emotion-related words. Word stimulations can induce different emotional and physiological responses (25, 26) and might influence acoustic differences between the depression group and control group (6). However, it is not clear whether the different emotional stimuli could play a role in the discriminant analysis, so we explored this further. After reading a paragraph, the participants were asked to rest for 1 min before moving on to the next paragraph. Voices were analyzed with OpenSMILE v2.1.0 (27) using the feature set of INTERSPEECH 2009 Emotion Challenge pre-set configuration (IS09_emotion.conf). The mean values of the following acoustic features within the utterance segments were calculated for the 12 dimensions of MFCC, ZCR, HNR, and F0.

Statistical Analyses

All statistical analyses were performed using the SPSS 21.0. For normally distributed continuous variables, the independent sample t-test was used for comparison between groups. Categorical data were analyzed using the chi-square test. Multivariate analysis of variance was used to compare the differences in acoustic characteristics between groups, with gender, age, and education level as covariates. To address the problem of multiple comparisons (15 acoustic features), a false discovery rate (FDR) procedure was further performed at a q-value of 0.05 over all acoustic features together. Spearman correlation analysis was performed to explore the relationship between depression symptom scores and acoustic characteristics. Further, multiple linear regression was conducted to explore the relation between clinical symptom and acoustic characteristics.

To choose the more suitable discriminant model, the discriminant effect on the depression group and normal control group was tested via logistic regression (LR) and support vector machine (SVM) models, respectively, both of which were used frequently in previous studies. Due to its superior discriminant effect, the LR model was selected for further analysis. All machine learning processes were completed with scikit-learn in Python 3.8. All parameters were performed by default to avoid the influence of manual intervention on discrimination results. The f_classif method in the feature_selection module of scikit-learn was used to screen all of acoustic features. The LR analysis was performed again based on the 20 selected variables, while analyzing the contribution of each variable to the classification. We performed a 10-fold cross validation of the classification and divided the dataset into the training set and the test set at a 7:3 ratio. The test data was only used for the validation of the discriminant model and were not used in any training process. Figures were generated from Matplotlib, a popular data visualization package for Python (38).

Results

General Demographic Data of the Two Groups

We enrolled 71 patients with depression and 62 healthy controls in Beijing, China. There were no significant differences in age, gender, or education between the two groups (Table 1). The depression group was accompanied by obvious symptoms of anxiety. The group’s psychic anxiety score of HAMA was 12.30 ± 5.12, while the somatic anxiety score of HAMA was 9.16 ± 6.00. The subscale scores of HAMD were 6.35 ± 3.13 (anxiety/somatization score), 0.57 ± 0.69 (weight score), 1.97 ± 1.44 (cognition score), 4.49 ± 2.02 (retardation score), and 2.86 ± 1.86 (insomnia score).

TABLE 1
www.frontiersin.org

Table 1. Demographic and clinical characteristics of all participants N = 133.

Multivariate Analysis of Variance of Acoustic Characteristics Between Depression and Normal Control Groups

A multivariate analysis of variance (ANOVA)—with gender, age, and education level as covariates—showed significant inter-group main effects of all of the acoustic characteristics (p < 0.0001), which survived FDR-correction. For η2 0.01, 0.06, and 0.14 were considered small, moderate, and large effect sizes, respectively. ZCR, F0, MFCC1-5, MFCC7-10 showed a large effect size, and the remaining features showed a moderate effect size. Only MFCC2, MFCC3, MFCC8, and MFCC9 had statistically significant conditional main effects in the positive, neutral, and negative emotion tasks (η2 = 0.06, 0.09, 0.05, and 0.02, respectively), but none of the acoustic characteristics had a group × condition interaction effect (Table 2).

TABLE 2
www.frontiersin.org

Table 2. Differences in acoustic characteristics between the depression and control groups under different emotional tasks N = 133.

Correlation Between Depressive Symptom Scores and Acoustic Characteristics of the Depression Group

Since there were significant differences between the groups for each acoustic characteristic indicator, we performed a Spearman correlation analysis for all acoustic features and the PHQ-9, HAMA, and HAMD scores. The results showed that MFCC4 and MFCC7 had significant positive correlations with the PHQ-9 score, regardless of the positive, neutral, or negative emotion tasks. ZCR in the positive condition task showed a significant correlation with the HAMA total score (r = 0.31) and with the HAMA somatic anxiety subscale score (r = 0.34). MFCC9 in the neutral task showed a negative correlation with the anxiety/somatization subscale score of the HAMD (r = −0.34). The remaining acoustic characteristics did not show a significant correlation with depressive symptoms (Table 3).

TABLE 3
www.frontiersin.org

Table 3. The correlation between depressive symptom severity and acoustic characteristics.

Linear Regression Analysis Between Clinical Symptom Scores and Acoustic Features

In linear regression, the PHQ-9 score was a dependent variable, while age, gender, education, MFCC4, and MFCC7 were independent variables. To avoid high collinearity of the same features under different emotion tasks, we selected MFCC4-neutral and MFCC7-negative, which had the highest correlation coefficient with the PHQ-9 score. Stepwise method was used. The result showed statistically significant effects, F = 7.25, R2 = 0.20, p = 0.01 (Table 4).

TABLE 4
www.frontiersin.org

Table 4. Linear regression analysis between PHQ-9 score and MFCC7-negative.

Then, the HAMD anxiety/somatization score was used as a dependent variable, while MFCC9-neutral, age, gender, and education were independent variables. Similarly, stepwise method was used. Here too, linear regression showed statistically significant effects, F = 4.15, R2 = 0.11, p = 0.049 (Table 4).

Discriminant Analysis of Acoustic Characteristics of the Depression Group

The LR model showed better discriminant results than SVM between the depression and control groups, with a discrimination accuracy of 89.66% and AUC of 0.98. A recall of 0.94 signified that 94% of depression patients were predicted as the depressed group, and precision of 0.89 signified that 89% of participants in the predicted depressed group were actually depression patients. Further, 83% of the healthy controls were predicted to be in the control group, and 91% of the participants in the predicted control group were actually healthy controls. The discrimination accuracy of the SVM model and the AUC were 86.21% and 0.95, respectively (see Table 5 and Figure 1).

TABLE 5
www.frontiersin.org

Table 5. The discriminant result of grouping by different classifiers.

FIGURE 1
www.frontiersin.org

Figure 1. ROC curves of acoustic features in LR and SVM models. The red line represents LR model; the green line represents SVM model; the horizontal axis is the false positive rate, and the vertical axis is the true positive rate.

Figure 2 shows the 20 acoustic features with the largest contribution to discrimination in the LR model. Concerning acoustic features—particularly MFCC1-negative, MFCC6-neutral—with contributions greater than 0, the larger the value, the more likely the participant is to be identified as depressed. In contrast, when these acoustic features have contributions less than 0, the smaller the value, the more likely the participant is to be identified as depressed.

FIGURE 2
www.frontiersin.org

Figure 2. The 20 acoustic features with the largest contribution in LR model. pos, positive emotion task; neg, negative emotion task; neu, neutral emotion task.

Discussion

There were significant differences in acoustic characteristics between the depression and normal control groups, after controlling for gender, age, and education. Only MFCC2, MFCC3, MFCC8, and MFCC9 showed differences during the positive, neutral, and negative emotion tasks. This suggests that there are differences in some acoustic features under different implicit emotional task patterns. Taguchi et al. also found that MFCC2 was higher in depressed patients than in normal controls, which is consistent with our results, but they did not report an association between MFCC and depressive symptom severity (5). These inconsistent results may be due to differences in language, disease stages, and speech recording methods. Another study from China found that MFCC5 and MFCC7 of the depression group were consistently lower than those of the control group across different emotional and situational tasks, and F0 was lower while ZCR was higher for the depression group in the reading text task, which was contrary to the findings of the present study (6). This inconsistency may be attributed to the different clinical characteristics of depression, such as anxiety and psychomotor retardation. The accuracy of identifying depression using acoustic characteristics was close to 90%. These results suggest that acoustic features may comprise a useful biomarker for diagnosing depression. Moreover, ZCR, MFCC4, MFCC7, and MFCC9 were associated with depressive and anxiety symptoms severity. MFCC7-negative can predict a PHQ-9 score; meanwhile, MFCC9-neutral can predict an anxiety/somatization score. Therefore, we believed that MFCC7 and MFCC9 can partly reflect the severity of depressive and anxiety symptoms and have the potential to be quantitative, objective indicators in evaluating disease severity.

Patients with depression often have autonomic nervous function disorders, and the muscle tone of their vocal tract can be affected by the disease, which may be one of the factors that distinguish depressed patients’ acoustic characteristics from those of healthy individuals. Patients with depression often feel tired, and a study suggests that sympathetic nervous system activity is enhanced, and parasympathetic nervous system activity is weakened during fatigue (28). Changes in the autonomic nervous system cause disturbances in the respiratory rate (29), and this may affect the airflow from the lungs to the vocal tract. Muscle tension is also a common symptom of depression (30). Changes in muscle tone can alter the dynamics of the vocal tract, limiting joint movement (31), which in turn slows speech and affects sound frequency. Air passes through the vocal cords and glottis, and the vocal tract acts as a filter that amplifies and attenuates different sound frequencies through movement, adjusting the spectrum of the glottis waveform (32). In a previous study, MFCC was associated with muscle movement of the vocal tract on MRI scans (16); MFCC can also reflect voice mood and is minimally affected by age and sex (5). We found that ZCR, MFCC4, MFCC7, and MFCC9 were associated with depression symptom severity. However, how depression affects muscle movement of the vocal tract that causes such changes in the specific frequencies of the patient’s voice remains unknown.

Speakers can constantly monitor and adjust their voice by using their speech as feedback in the auditory circuit (32). In the central auditory pathway, 5-hydroxytryptamine (5-HT) exists in many structures from the cochlear nucleus to the auditory cortex. It is one of the most important neurotransmitters in auditory processing; 5-HT neurons in the inferior colliculus can adjust the frequency of perceived tones by changing the response latency of neurons (33). Female patients with depression show reduced tolerance to sound, also known as auditory hypersensitivity (34). People with low expression of 5-HT transporter perceived sounds as clearer than those with high expression, showing a higher signal-to-noise ratio and volume (35). Therefore, patients with depression may also be more sensitive than healthy individuals when hearing the sounds of their own voices. Due to the low 5-HT function, the perceived frequency of their own sounds may change. Through auditory feedback, patients may tend to have lower intonation and acoustic features, which are different from those of normal controls when speaking.

Rats express their emotions through ultrasonic vocalizations, and different emotions correspond to different frequencies of ultrasound; ultrasonic vocalizations in anxious conditions due to isolation are associated with 5-HT. Drugs that act on the 5-HT receptor, such as some serotonin reuptake inhibitors and serotonin 5-HT1A receptor agonists, dose-dependently reduce isolation-induced ultrasound vocalizations in rats (36). Serotonin reuptake transporter gene-deficient rats exhibited more anxious behavior, accompanied by changes in the ultrasonic vocal structure, and they were less likely to make orectic 50 kHz vocalizations when exploring novel environments (37). The above studies on animals suggest that 5-HT may be related to the frequency and behavior of the sound emitted. However, there are no studies on human vocalization and 5-HT. Given that 5-HT plays an important role in the pathogenesis of depression, our results lead us to suspect that 5-HT may be involved in the pathological mechanism of the abnormal acoustic characteristics of depressed patients.

Limitations

Although the findings of this study have implications for improving the diagnosis of depression, there are some inherent limitations to the study design. First, the exploratory cross-sectional design of this study precluded causal inferences. Longitudinal follow-ups of patients with depression, including observation of the entire course of disease with and without drug treatment, need to be carried out in the future. Second, the clinical symptom assessment in this study was simple; future studies should include more dimensions of depression, including the assessment of suicide risk and autonomic nervous function. Third, the experimental paradigm of this study was relatively simple and warrants further expansion to simulate different scenarios, help gain a comprehensive understanding of the changes in the acoustic characteristics of depressed patients, and fully evaluate their accuracy as objective biomarkers of depression. Furthermore, this study only included Chinese Mandarin samples, whether the specific language or other cultural factors may influence the generalizability of the results is still unclear. Finally, although we performed a 10-fold cross validation, a larger sample size and independent validation datasets may be more effective for verifying our conclusions.

Conclusion

The change in acoustic characteristics in patients with depression may not be caused by a single factor, and the specific underlying mechanism remains unclear. However, the acoustic characteristics of depressed patients were significantly different from those of the normal control group in this study, and the accuracy of depression classification based on acoustic features was very high. These findings suggest that acoustic characteristics may become objective biomarkers of depression; this may help clinicians make objective diagnosis of depression, reduce the rate of missed diagnosis and misdiagnosis, and promote the development of Internet medical treatment.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

The studies involving human participants were reviewed and approved by the Ethics Committee of Beijing Huilongguan Hospital. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

S-PT, Z-RW, and Y-LT contributed to the conception and design of the study. QZ performed the statistical analysis and wrote the first draft of the manuscript. S-PT reviewed and revised the manuscript. H-ZF, Y-LZ, and Z-XT organized the database. Y-LL, LL, and Y-XW contributed to provision of study materials and patients. All authors contributed to the manuscript revision, as well as read and approved the submitted version.

Funding

This work was funded by the Funds of Beijing Science and Technology Project (code: Z191100006619104) and the Natural Science Foundation of Beijing (code: 7202072). The funding sources of this study had no role in the study design, data collection, and analysis, decision to publish, or preparation of the article.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We express our gratitude to all participants of this study. We would like to thank Editage (www.editage.cn) for English language editing.

References

1. Huang Y, Wang Y, Wang H, Liu Z, Yu X, Yan J, et al. Prevalence of mental disorders in China: a cross-sectional epidemiological study. Lancet Psychiatry. (2019) 6:211–24.

Google Scholar

2. Mitchell AJ, Rao S, Vaze A. Can general practitioners identify people with distress and mild depression? A meta-analysis of clinical accuracy. J Affect Disord. (2011) 130:26–36. doi: 10.1016/j.jad.2010.07.028

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Lam RW, McIntosh D, Wang J, Enns MW, Kolivakis T, Michalak EE, et al. Canadian network for mood and anxiety treatments (CANMAT) 2016 clinical guidelines for the management of adults with major depressive disorder: section 1. disease burden and principles of care. Can J Psychiatry. (2016) 61:510–23. doi: 10.1177/0706743716659416

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Liu Q, He H, Yang J, Feng X, Zhao F, Lyu J. Changes in the global burden of depression from 1990 to 2017: findings from the global burden of disease study. J Psychiatr Res. (2020) 126:134–40. doi: 10.1016/j.jpsychires.2019.08.002

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Taguchi T, Tachikawa H, Nemoto K, Suzuki M, Nagano T, Tachibana R, et al. Major depressive disorder discrimination using vocal acoustic features. J Affect Disord. (2018) 225:214–20. doi: 10.1016/j.jad.2017.08.038

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Wang J, Zhang L, Liu T, Pan W, Hu B, Zhu T. Acoustic differences between healthy and depressed people: a cross-situation study. BMC Psychiatry. (2019) 19:300. doi: 10.1186/s12888-019-2300-7

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Krajewski J, Schnieder S, Sommer D, Batliner A, Schuller B. Applying multiple classifiers and non-linear dynamics features for detecting sleepiness from speech. Neurocomputing. (2012) 84:65–75.5.

Google Scholar

8. Grabowski K, Rynkiewicz A, Lassalle A, Baron-Cohen S, Schuller B, Cummins N, et al. Emotional expression in psychiatric conditions: new technology for clinicians. Psychiatry Clin Neurosci. (2019) 73:50–62. doi: 10.1111/pcn.12799

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Tolkmitt FJ, Scherer KR. Effect of experimentally induced stress on vocal parameters. J Exp Psychol Hum Percept Perform. (1986) 12:302–13. doi: 10.1037//0096-1523.12.3.302

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Schirmer A, Adolphs R. Emotion perception from face, voice, and touch: comparisons and convergence. Trends Cogn Sci. (2017) 21:216–28. doi: 10.1016/j.tics.2017.01.001

PubMed Abstract | CrossRef Full Text | Google Scholar

11. France DJ, Shiavi RG, Silverman S, Silverman M, Wilkes DM. Acoustical properties of speech as indicators of depression and suicidal risk. IEEE Trans Biomed Eng. (2000) 47:829–37. doi: 10.1109/10.846676

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Alpert M, Pouget ER, Silva RR. Reflections of depression in acoustic measures of the patient’s speech. J Affect Disord. (2001) 66:59–69. doi: 10.1016/s0165-0327(00)00335-9

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Cannizzaro M, Harel B, Reilly N, Chappell P, Snyder PJ. Voice acoustical measurement of the severity of major depression. Brain Cogn. (2004) 56:30–5. doi: 10.1016/j.bandc.2004.05.003

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Mundt JC, Snyder PJ, Cannizzaro MS, Chappie K, Geralts DS. Voice acoustic measures of depression severity and treatment response collected via interactive voice response (IVR) technology. J Neurolinguist. (2007) 20:50–64. doi: 10.1016/j.jneuroling.2006.04.001

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Li H, Xu X, Wu G, et al. Research on speech emotion feature extraction based on MFCC. J Electr Measur Instrument. (2017) 31: 448–53.

Google Scholar

16. Zhu Y, Kim YC, Proctor MI, Narayanan SS, Nayak KS. Dynamic 3-D visualization of vocal tract shaping during speech. IEEE Trans Med Imaging. (2013) 32:838–48. doi: 10.1109/TMI.2012.2230017

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Ozdas A, Shiavi RG, Wilkes DM, Silverman MK, Silverman SE. Analysis of vocal tract characteristics for near-term suicidal risk assessment. Methods Inf Med. (2004) 43:36–8. doi: 10.1055/s-0038-1633420

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Akkaralaertsest T, Yingthawornsuk T. Comparative analysis of vocal characteristics in speakers with depression and high-risk suicide[J]. Int J Comp Theory Engine. (2015) 7:448. doi: 10.7763/ijcte.2015.v7.1001

CrossRef Full Text | Google Scholar

19. Mundt JC, Vogel AP, Feltner DE, Lenderking WR. Vocal acoustic biomarkers of depression severity and treatment response. Biol Psychiatry. (2012) 72:580–7. doi: 10.1016/j.biopsych.2012.03.015

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Long H, Guo Z, Wu X, Hu B, Liu Z, et al. Detecting depression in speech: comparison and combination between different speech types[C]//2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE. (2017) 2017:1052–8.

Google Scholar

21. Hamilton M. A rating scale for depression. J Neurol Neurosurg Psychiatry. (1960) 23:56–62.

Google Scholar

22. Zhao B, Li Z, Wang Y, Ma X, Wang X, Wang X, et al. Can acupuncture combined with SSRIs improve clinical symptoms and quality of life in patients with depression? Secondary outcomes of a pragmatic randomized controlled trial. Complement Ther Med. (2019) 45:295–302. doi: 10.1016/j.ctim.2019.03.015

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Hamilton M. The assessment of anxiety states by rating. Br J Med Psychol. (1959) 32:50–5. doi: 10.1111/j.2044-8341.1959.tb00467.x

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Löwe B, Unützer J, Callahan CM, Perkins AJ, Kroenke K. Monitoring depression treatment outcomes with the patient health questionnaire-9. Med Care. (2004) 42:1194–201. doi: 10.1097/00005650-200412000-00006

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Wu D, Yao S, Guo W. Difference ERPs effects of the difference introduction on the recognition of Chinese emotional content words in healthy subjects[C]//2005 International Conference on Neural Networks and Brain. IEEE (2005) 2:1303–6.

Google Scholar

26. Xu S, Yin H, Wu D. Initial establishment of the Chinese Affective Words Categorize System used in research of emotional disorder[J]. Chin. Mental Health J. (2008) 22:770–4.

Google Scholar

27. Eyben F, Weninger F, Gross F, Schuller B. Recent developments in openSMILE, the munich open-source multimedia feature extractor. MM 2013 - Proceedings of the 2013 ACM Multimedia Conference. New York, NY: ACM (2013).

Google Scholar

28. Tanaka M, Tajima S, Mizuno K, Ishii A, Konishi Y, Miike T, et al. Frontier studies on fatigue, autonomic nerve dysfunction, and sleep-rhythm disorder. J Physiol Sci. (2015) 65:483–98. doi: 10.1007/s12576-015-0399-y

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Kreibig SD. Autonomic nervous system activity in emotion: a review. Biol Psychol. (2010) 84:394–421. doi: 10.1016/j.biopsycho.2010.03.010

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Maes M. “Functional” or “psychosomatic” symptoms, e.g. a flu-like malaise, aches and pain and fatigue, are major features of major and in particular of melancholic depression. Neuro Endocrinol Lett. (2009) 30:564–73.

PubMed Abstract | Google Scholar

31. Roy N, Nissen SL, Dromey C, Sapir S. Articulatory changes in muscle tension dysphonia: evidence of vowel space expansion following manual circumlaryngeal therapy. J Commun Disord. (2009) 42:124–35. doi: 10.1016/j.jcomdis.2008.10.001

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Postma A. Detection of errors during speech production: a review of speech monitoring models. Cognition. (2000) 77:97–132. doi: 10.1016/s0010-0277(00)00090-1

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Hurley LM, Thompson AM, Pollak GD. Serotonin in the inferior colliculus[J]. Hearing Res. (2002) 168:1–11.

Google Scholar

34. Gopal KV, Briley KA, Goodale ES, Hendea OM. Selective serotonin reuptake inhibitors treatment effects on auditory measures in depressed female subjects. Eur J Pharmacol. (2005) 520:59–69. doi: 10.1016/j.ejphar.2005.07.019

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Selinger L, Zarnowiec K, Via M, Clemente IC, Escera C. Involvement of the Serotonin Transporter Gene in Accurate Subcortical Speech Encoding. J Neurosci. (2016) 36:10782–90. doi: 10.1523/JNEUROSCI.1595-16.2016

PubMed Abstract | CrossRef Full Text | Google Scholar

36. Wöhr M, van Gaalen MM, Schwarting RK. Affective communication in rodents: serotonin and its modulating role in ultrasonic vocalizations. Behav Pharmacol. (2015) 26:506–21. doi: 10.1097/FBP.0000000000000172

PubMed Abstract | CrossRef Full Text | Google Scholar

37. Golebiowska J, Hołuj M, Potasiewicz A, Piotrowska D, Kuziak A, Popik P, et al. Serotonin transporter deficiency alters socioemotional ultrasonic communication in rats. Sci Rep. (2019) 9:20283. doi: 10.1038/s41598-019-56629-y

PubMed Abstract | CrossRef Full Text | Google Scholar

38. Hunter, J. D. Matplotlib: a 2D graphics environment. Comput. Sci. Eng. (2007) 9:90–5. doi: 10.1109/MCSE.2007.55

CrossRef Full Text | Google Scholar

Keywords: depression, acoustic characteristics, MFCC, biomarker, zero-crossing rate

Citation: Zhao Q, Fan H-Z, Li Y-L, Liu L, Wu Y-X, Zhao Y-L, Tian Z-X, Wang Z-R, Tan Y-L and Tan S-P (2022) Vocal Acoustic Features as Potential Biomarkers for Identifying/Diagnosing Depression: A Cross-Sectional Study. Front. Psychiatry 13:815678. doi: 10.3389/fpsyt.2022.815678

Received: 15 November 2021; Accepted: 30 March 2022;
Published: 28 April 2022.

Edited by:

Takahiro A. Kato, Kyushu University, Japan

Reviewed by:

Bingxin Zhao, Purdue University, United States
Tingshao Zhu, Institute of Psychology (CAS), China

Copyright © 2022 Zhao, Fan, Li, Liu, Wu, Zhao, Tian, Wang, Tan and Tan. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Shu-Ping Tan, shupingtan@126.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.