ORIGINAL RESEARCH article

Front. Psychiatry, 20 July 2023

Sec. Digital Mental Health

Volume 14 - 2023 | https://doi.org/10.3389/fpsyt.2023.1079448

Exploring the ability of vocal biomarkers in distinguishing depression from bipolar disorder, schizophrenia, and healthy controls

  • 1. Key Laboratory of Adolescent Cyberpsychology and Behavior (CCNU), Ministry of Education, Wuhan, China

  • 2. School of Psychology, Central China Normal University, Wuhan, China

  • 3. Key Laboratory of Human Development and Mental Health of Hubei Province, Wuhan, China

  • 4. Wuhan Wuchang Hospital, Wuchang Hospital Affiliated to Wuhan University of Science and Technology, Wuhan, China

  • 5. Institute of Psychology, Chinese Academy of Sciences, Beijing, China

  • 6. CAS Key Laboratory of Behavioral Science, Institute of Psychology, Chinese Academy of Sciences, Beijing, China

Article metrics

View details

11

Citations

4,1k

Views

1,5k

Downloads

Abstract

Background:

Vocal features have been exploited to distinguish depression from healthy controls. While there have been some claims for success, the degree to which changes in vocal features are specific to depression has not been systematically studied. Hence, we examined the performances of vocal features in differentiating depression from bipolar disorder (BD), schizophrenia and healthy controls, as well as pairwise classifications for the three disorders.

Methods:

We sampled 32 bipolar disorder patients, 106 depression patients, 114 healthy controls, and 20 schizophrenia patients. We extracted i-vectors from Mel-frequency cepstrum coefficients (MFCCs), and built logistic regression models with ridge regularization and 5-fold cross-validation on the training set, then applied models to the test set. There were seven classification tasks: any disorder versus healthy controls; depression versus healthy controls; BD versus healthy controls; schizophrenia versus healthy controls; depression versus BD; depression versus schizophrenia; BD versus schizophrenia.

Results:

The area under curve (AUC) score for classifying depression and bipolar disorder was 0.5 (F-score = 0.44). For other comparisons, the AUC scores ranged from 0.75 to 0.92, and the F-scores ranged from 0.73 to 0.91. The model performance (AUC) of classifying depression and bipolar disorder was significantly worse than that of classifying bipolar disorder and schizophrenia (corrected p < 0.05). While there were no significant differences in the remaining pairwise comparisons of the 7 classification tasks.

Conclusion:

Vocal features showed discriminatory potential in classifying depression and the healthy controls, as well as between depression and other mental disorders. Future research should systematically examine the mechanisms of voice features in distinguishing depression with other mental disorders and develop more sophisticated machine learning models so that voice can assist clinical diagnosis better.

Introduction

The identification and diagnosis of depression through clinical interviews are often slow and unreliable (14). About half of cases go unrecognized: in a meta-analysis of 41 studies, recognition accuracy of depression by general practitioners was 47.3% (5). Therefore, accurate and fast ways to identify cases of depression will have major clinical benefits.

Novel applications of computational methods are making some inroads into this problem. For example, a review of 14 studies indicated that both sensitivity and specificity of diagnostic performance of deep learning models were higher than that of health-care professionals (6). In the last decade, there have been interests in the ability of exploiting vocal biomarkers to identify depression with machine learning methods to investigate whether voice can be used as an auxiliary tool to assist clinical diagnosis (711). Previous research mainly focused on examining the ability of vocal features in classifying individuals with depression and healthy population (1217), and F-measure of relevant classifiers reached 0.9 (12). These findings suggest that vocal biomarkers may have discriminatory potential in identifying depression. While the differential diagnosis is complicated by the presence of other mental disorders.

The prevalence of mental disorders, according to a study in China, showed that the weighted prevalence of any disorder (excluding dementia) was 16.6% (95% CI 13.0–20.2) during the 12 months before the interview (18). Clinically, the psychiatric diagnosis necessitates distinguishing depressed individuals not only from the healthy ones, but also from other mental illnesses with similar mood symptoms or similar voice patterns. It needs to be stressed that voice conveys emotion related information. In the field of affective computing, voice features were shown their abilities in recognizing different kinds of emotions (19, 20). Both bipolar disorder (BD) and schizophrenia exhibit symptoms comparable to depression. According to the Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition (DSM-4) (21), BD involves both depressive and manic episodes, while schizophrenia patients with negative symptoms report anhedonia (loss of ability to experience pleasure). The specificity of alterations in vocal characteristics to depression has not been rigorously investigated.

Variations in vocal features have been observed in schizophrenia and BD patients. Acoustic analyses of speech in schizophrenia have revealed subtle aberrancies in pitch variability associated with flat affect, as well as more pronounced deviations in properties such as percentage of speaking time, speech rate, and pause duration that correlate with symptoms of alogia and blunted affect (7). Espinola and her colleagues (22) constructed classification models based on speech samples from 31 individuals (20 of whom had a prior diagnosis of schizophrenia and 11 healthy controls). The classifiers attained an accuracy of 91.76% in distinguishing between groups. Regarding BD, studies have utilized vocal features to predict patients’ emotional states (e.g., depressive, manic, mixed states). Classification analyses yielded an area under curve (AUC) of 0.89 (9, 10). Another study investigated whether vocal features acquired via verbal fluency tasks could accurately differentiate mixed states in BD using machine learning methods. And results showed that for depressive versus mixed depressive episodes, the F-measure was 0.86, while for hypomanic versus mixed hypomanic episodes, the F-measure was 0.75 (23). These studies showed that voice features may also be informative for other psychiatric diagnosis. It should be noted that several studies have examined the utility of vocal features in developing classifiers for several mental disorders. A study (24) employed polytomous logistic regression analysis of vocal features to discriminate among healthy controls (n = 23), individuals with bipolar disorder (n = 8), and those with major depressive disorder (n = 14). The model attained 90.79% accuracy in classifying participants into the three diagnostic groups. Another study (25) proposed a methodology to support the diagnosis of several mental disorders using vocal acoustic analysis and machine learning. The results showed that random forests with 300 trees achieved the best classification performance (75.27% for accuracy) for the simultaneous detection of major depressive disorder (26), schizophrenia (20), BD (15), generalized anxiety disorder (4), and healthy controls (12). However, the datasets of above two studies were imbalanced for each group. The imbalanced dataset problems become more complicated in multi-class imbalanced classification tasks, in which there may be multiple minority and majority classes that cause skewed data distribution. And machine learning algorithms tend to favor the majority class samples, hence damaging the multi-classification results (27, 28). Moreover, extensive comparisons between mental disorders and healthy controls may offer more information about the effectiveness of voice for clinically complex differential diagnosis.

Various speech features are indicative of depression. Mel-frequency cepstrum coefficients (MFCCs) constitute the most prevalent vocal features employed in speech recognition systems and psychiatric condition classification models (26). MFCCs are obtained by extracting frequency spectral features of the speech signal using the short-time power spectrum, mapping these features onto the Mel scale to better present auditory characteristics, and then obtaining MFCC coefficients through cepstrum analysis that can characterize the speech envelope (12, 29). Multiple studies have demonstrated the utility of MFCCs in developing classification models for depression detection (3032). For example, Di et al. (29) employed MFCCs to classify major depression patients and healthy individuals, area under curve (AUC) reached 0.8.

The identity vector (i-vector) approach, grounded within the total variability framework, represents the state-of-the-art technique for speaker verification (12, 29, 32, 33). The total variability framework offers an effective means of capturing speaker- and channel-related variability in a low dimensional subspace. i-vectors are highly informative for encoding cepstral variability. Classification models based on i-vectors demonstrated capacity for identifying depression with high accuracy. For instance, prior work found i-vector based model outperformed a baseline model defined by KL-means supervectors (32). Nasir and his colleagues (33) used i-vectors to investigate various audio and visual features for classification, reporting high accuracy with i-vector modeling of MFCC features. Indeed, one study demonstrated a 40% improvement in predictive accuracy (F-score) with the i-vector methodology (12). And Di et al. (29) observed a 14% enhancement in model performance (AUC) with i-vectors relative to MFCCs alone. Although the participant cohorts were exclusively female in both studies, the results demonstrated the promise of i-vectors for enhancing the accuracy of machine learning models for depression classification.

The objective of this study was to evaluate the efficacy of vocal features as differential diagnostic markers for depression compared to other psychiatric disorders. Three binary classification paradigms were employed in total: (1) the capacity of voice features to distinguish any psychiatric condition (depression, bipolar disorder, schizophrenia) versus healthy controls at baseline; (2) the ability of voice features to differentiate a specific psychiatric illness from healthy controls; (3) the capability of vocal features to distinguish between discrete psychiatric disorders in a pairwise manner. Among these paradigms, the first one was served as a baseline to determine whether the dataset achieved performance commensurate with existing research, as well as a benchmark for model performance under other framework conditions. The second and third paradigms were employed to systematically evaluate the capacity of vocal characteristics to distinguish between case and control groups.

Methods

Participants

All participants were randomly recruited. All participants were Chinese aged 18 to 59 years. A diagnosis of primary psychiatric illness was established for all patients using the Diagnostic and Statistical Manual of Mental Disorders (DSM-IV) (21) by psychiatrists. Participants clustered into four categories based on diagnosis: healthy controls, depression, BD and schizophrenia. Clinical staging was further specified: depression subjects were actively symptomatic, BD patients were euthymic, and schizophrenia patients were in remission. Healthy controls were openly recruited. Patients with comorbid psychiatric conditions were excluded for all diagnostic categories. And the general exclusion criteria across all participants were: physical illnesses, pregnancy and lactating, substance abuse within 12 months. Demographic variables were age, gender and education level. See Table 1.

Table 1

GroupsGenderAgeEducationOccupation
Health58 males34.88 ± 10.547.78 ± 2.604.13 ± 3.48
57 females34.73 ± 9.697.47 ± 2.533.08 ± 3.13
Depression53 males32.67 ± 8.626.70 ± 2.362.34 ± 2.76
70 females34.31 ± 11.227.69 ± 2.253.32 ± 3.20
Bipolar16 males30.06 ± 10.466.75 ± 2.612.55 ± 2.74
21 females33.10 ± 10.597.02 ± 2.164.93 ± 3.14
Schizophrenia10 males27.86 ± 7.146.81 ± 2.333.84 ± 3.34
10 females30.81 ± 7.616.41 ± 2.284.64 ± 3.41

Demographic information about each group.

Measures

Four vocal tasks were employed for data collection: video watching (VW), text reading (TR), question answering (QA), picture description (PD). Each task incorporated positive, negative, neutral emotional primes to comprehensively represent existing research paradigms (3438). In VW, participants viewed video clips then described the most memorable scenes or figures. For QA, participants provided spoken answers to nine questions (three questions/emotion), e.g., “Please share your most wonderful experience and describe it in detail”. In TR, participants read three 140-word paragraphs aloud. For PD, participants described facial expressions and image content from the Chinese Facial Affective Picture System and the Chinese Affective Picture System (three facial affective pictures and three affective pictures for three emotion primes), respectively. Twenty one voice recordings were collected from each participant. The emotional priming effects of these tasks were validated in previous research (37, 38). Research also indicates that this dataset affords stable prediction accuracies across emotions and tasks (15, 35, 37, 38).

All participants were seated 1 m from a 21-inch monitor. Instructions were displayed on-screen. Speech was recorded 50 cm distant using a professional condenser microphone (Neumann TLM102, Germany) and a voice recorder (RME Fireface UCX, Germany). The experimenter controlled recording initiation and termination for each participant to exclude the experimenter’s speech from recordings. Participants were asked to complete all tasks in random order. Ambient noise was under 60 dB. Recordings less than 10 s were excluded. Recording duration details were displayed in Table 2. The speech was Mandarin Chinese. Recordings were collected with a sampling rate of 44.1 kHz and 24-bit. Informed consent was obtained in writing pre-experiment. This study was part of a national project and was approved by the Institutional Review Board (IRB) of Institute of Psychology, Chinese Academy of Sciences.

Table 2

GroupsMinMaxM ± SD
Health1018827.8 ± 17.99
Depression1016429.4 ± 17.59
Bipolar1015630.77 ± 17.74
Schizophrenia1014930.15 ± 20.83

Duration descriptions of voice recordings in each group(s).

Data analysis

Preprocessing

A total of seven classification tasks were employed: any disorder versus healthy controls; depression versus healthy controls; BD versus healthy controls; schizophrenia versus healthy controls; depression versus BD; depression versus schizophrenia; BD versus schizophrenia (AH, DH, BH, SH, DB, DS, BS). In each task, either the mental disorder group or depression group was designated as the case group, with the other constituting the control group.

For model building, the data were randomly split into training set (70%) and test set (30%). Given the small sample size, group differences in demographics were assessed using permutation test (a permutation t test), a nonparametric method suitable for small samples with unknown distribution. Many parametric tests have their corresponding permutation test versions employing the same test statistic but deriving p-values from the sample-specific permutation distribution of that statistic, rather than from the theoretical distribution derived from the parametric assumption (39).

Previous research indicates demographic factors confound the detection of depression from voice features (15). It was therefore critical to control for demographics to isolate the role of vocal features. Age, gender and education differences between groups were evaluated in the training set. Owing to the small, uneven group sizes, matching was performed twice: (1) matched gender within categories through exact matching under the guidance of random sampling (a coarsened matching method for binary variables) (40) to match the number of males and females according to the smaller number within the group by the sample function in R (41). For example, 16 females were randomly selected in the BD group. After matching, there were 32 BD patients; 106 depression patients, 114 healthy ones, and 20 schizophrenia patients with unbiased sex ratios in each category. (2) Case-control matching within each classification task, e.g., 32 BD patients and 32 healthy controls in the model of classifying the BD versus healthy control task. Ultimately, the demographics of cases and controls were preliminarily balanced for each task.

Mel frequency cepstral coefficients

Mel Frequency Cepstral Coefficients (MFCCs) were extracted by MATLAB R2020b (42) with a window size of 25 ms, a window shift of 10 ms, a pre-emphasis filter with coefficient 0.97, and a sinusoidal lifter with coefficient 22 (12). A filter bank with 23 filters was used and 13 coefficients were extracted. Utterances were downsampled to 8 kHz before feature extraction. The first and second derivatives of MFCCs were also extracted.

MFCCs were extracted for each vocal task from every participant. For each participant, the mean values across the 21 tasks were calculated to streamline modeling. And prior research validates the consistent efficacy of vocal features across vocal tasks (15, 35, 37, 38).

I-vector extraction

The i-vector extraction formula is represented as follows:

where m is the mean super-vector of the Universal Background Model (UBM). UBM representing the feature distribution of the acoustic space, is adapted to a set of given speech frames to estimate utterance-dependent Gaussian Mixture Models parameters. M is the mean-centered super-vector of the speech utterance derived using the 0th and 1st order Baum–Welch statistics. v is the i-vector, the representation of a speech utterance (43).

Twenty i-vectors were derived for each participant. All i-vectors underwent quantile normalization. The training data was then used for constructing logistic regression models. i-vectors extraction was performed using the Kaldi toolkit (44).

Logistic regression

i-vectors were subjected to logistic regression on the training set using R (41). Logistic regression employed the Glmnet method (45) with ridge regularization and 5-fold cross validation. The resultant models were then applied to the test sets for each classification task.

Model building

To evaluate the classification ability of voice features for differential diagnosis, logistic regression models were constructed for the seven binary classification tasks. First, we examined the classification ability of voice features in distinguishing between the healthy controls and those with any mental disorder. Second, we examined model performances in separately classifying the healthy controls and each discrete clinical group. Third, model performances were evaluated in pairwise classification among the three disorders (depression versus BD; depression versus schizophrenia; BD versus schizophrenia).

Results

Descriptive statistics

Age differences were tested. As shown in Table 3, cases and controls for all classification tasks were matched on relevant variables, except for the schizophrenia versus healthy control task. Propensity score matching (46) was therefore conducted to re-match the schizophrenia and the healthy control group. The case group and control group for this task were then balanced, as detailed in Table 4.

Table 3

TasksGroupsM ± SDtp
AHaAny disorder32.35 ± 9.82−1.420.16
Healthy controls34.48 ± 10.63
DHDepression33.76 ± 10.23−0.40.69
Healthy controls34.43 ± 10.48
BHBD29.68 ± 8.99−1.850.07
Healthy controls34.86 ± 9.03
SHSchizophrenia29.07 ± 7.42−2.110.03*
Healthy controls37.14 ± 7.42
DBDepression34.18 ± 9.671.570.12
BD29.68 ± 8.99
DSDepression33.86 ± 9.491.450.15
Schizophrenia29.07 ± 7.42
BSBD30.00 ± 7.750.330.77
Schizophrenia29.07 ± 7.42

t-tests of age difference for each classification task.

a

AH, any disorder versus healthy controls; DH, depression versus healthy controls; BH, BD versus healthy controls; SH, schizophrenia versus healthy controls; DB, depression versus BD; DS, depression versus schizophrenia; BS, BD versus schizophrenia. Similarly hereinafter.

*p < 0.05.

Table 4

TasksGroupsM ± SDtp
S_HSchizophrenia28.79 ± 8.15−0.740.48
Healthy controls31.14 ± 8.74

t-test after propensity score matching for SH task.

Permutation tests were conducted for both gender and education. No significant differences either in gender or education across different tasks. See Tables 5, 6.

Table 5

TasksGroupsGenderϰ2p
MaleFemale
AHAny disorder555501
Healthy controls4040
DHDepression373701
Healthy controls3737
BHBD111101
Healthy controls1111
SHSchizophrenia7701
Healthy controls77
DBDepression111101
BD1111
DSDepression7701
Schizophrenia77
BSBD7701
Schizophrenia77

Permutation tests on gender for each classification task.

Table 6

TasksGroupsEducationϰ2p
High school level and belowUndergraduate and above
AHAny disorder46640.180.76
Healthy controls3149
DHDepression29450.031
Healthy controls2846
BHBD91301
Healthy controls913
SHSchizophrenia681.290.45
Healthy controls95
DBDepression10120.091
BD913
DSDepression591.290.45
Schizophrenia86
BSBD770.141
Schizophrenia86

Permutation tests on education for each classification task.

Following matching, duration differences between groups were evaluated for each classification task. The mean value of 21 recording durations was first computed for each participant across all experimental tasks. Difference tests were then conducted between groups for each classification task. Results showed that there were no significant differences between groups for any classification task. See Table 7.

Table 7

TasksGroupsM ± SDtp
AHAny disorder30.47 ± 8.841.530.13
Healthy controls28.65 ± 7.44
DHDepression30.62 ± 7.801.590.12
Healthy controls28.22 ± 10.36
BHBD30.09 ± 9.930.630.53
Healthy controls28.42 ± 7.81
SHSchizophrenia29.22 ± 6.660.140.89
Healthy controls28.82 ± 8.65
DBDepression30.94 ± 7.680.570.57
BD29.58 ± 8.17
DSDepression31.21 ± 7.220.770.45
Schizophrenia29.17 ± 6.60
BSBD29.54 ± 10.010.420.68
Schizophrenia28.22 ± 5.59

Duration differences between groups for each classification task.

Classification

Model performance metrics for different tasks were summarized in Table 8. For the general ability of vocal features to classify healthy versus any clinical group (AH task), the F-score was 0.82, AUC (area under curve) was 0.79. In distinguishing specific disorders from healthy controls, results showed: for DH task, F-score = 0.78, AUC = 0.77; for BH task, F-score = 0.80, AUC = 0.80; for SH task, F-score = 0.73, AUC = 0.75. To further examine the ability of voice features on pairwise classifications among the three mental disorders, DB, DS and BS tasks were performed. Results showed: for DB task F-score = 0.44, AUC = 0.50; for DS task, F-score = 0.83, AUC = 0.83; for BS task, F-score = 0.91, AUC = 0.92.

Table 8

TasksSensitivitySpecificityAccuracyPrecisionRecallF-scoreAUC
AH0.810.760.790.830.810.820.79
DH0.810.720.770.740.810.780.77
BH0.800.800.800.800.800.800.80
SH0.670.830.750.800.670.730.75
DB0.400.600.500.500.40.440.50
DS0.830.830.830.830.830.830.83
BS0.831.000.921.000.830.910.92

Results on 7 classification tasks with the i-vector framework.

We also compared model performances for all classification tasks. Following Bonferroni correction (n = 6), the only significant difference was that the depression versus BD model (DB task) showed worse performance (lower AUC) than the BD versus schizophrenia model (BS task; p < 0.05), as detailed in Table 9.

Table 9

Model comparisonsDeLong’s testdfpCorrected p
AH_DH0.32134.670.754.50
AH_BH−0.1129.020.915.46
AH_SH0.2713.780.794.74
AH_DB2.3225.540.030.18
AH_DS−0.3514.660.734.38
AH_BS−1.3418.710.201.20
DH_BH−0.3232.180.754.50
DH_SH0.1114.700.925.52
DH_DB2.0927.640.050.28
DH_DS−0.5215.880.613.66
DH_BS−1.5321.290.140.84
BH_SH0.3021.480.764.56
BH_DB2.0136.640.050.31
BH_DS−0.2223.920.834.98
BH_BS−0.9329.350.362.16
SH_DB1.4125.270.171.02
SH_DS−0.4721.630.653.90
SH_BS−1.0518.370.311.86
DB_DS−2.0227.560.050.32
DB_BS−2.9329.930.010.04*
DS_BS−0.5819.800.573.42

Pairwise comparisons of model performances for the 7 classification tasks.

*p < 0.05.

Discussion

In this research, we investigated the ability of vocal biomarkers to classify various health conditions. Following matching, descriptive statistics showed no differences in demographic variables between the case and control groups for any task, addressing potential threats to validity (15). Logistic regression models based on MFCC-derived i-vectors were developed for all classification tasks. Results indicated vocal features may assist differential diagnosis of depression, albeit with varying degrees of effectiveness across classification tasks.

The AH model examined the baseline ability of vocal features to distinguish clinical from healthy groups. Classifying the healthy group and the any-disorder group yielded an F-score of 0.82 and AUC of 0.79, demonstrating vocal features can distinguish mental disorders from the health.

Furthermore, the DH, BH, and SH models investigated the ability of voice features in distinguishing specific disorders from health. For DH, BH, and SH tasks, F-scores ranged from 0.73 to 0.80, AUC scores from 0.75 to 0.80. Model comparisons showed there were no significant differences in the pairwise comparisons among the AH and the three mental illnesses versus healthy control classification tasks. Our results about DH are consistent with the existing findings (8, 12, 15, 4750). For example, a previous study examined the significance of the association between voice features and depression using binary logistic regression, and the actual classification ability of voice features on depression using machine learning method, results showed that the contribution effect of voice features reached 35.65% (Nagelkerke’s R2), further classification model achieved 81% of F-measure (15). We set this classification model as another baseline for the reference of further classification. The results of BH and SH classification models are also consistent with existing studies. For instance, one study investigated whether voice features from naturalistic phone calls could discriminate between BD and healthy control individuals, results showed that compared to the control group, BD was classified with a 0.79 sensitivity and 0.76 AUC (51). For schizophrenia detection, Tahir et al. (52) classified schizophrenia patients and healthy controls with multilayer perceptron and the accuracy achieved 81.3%. It was also suggested that speech abnormalities, related to flat affect and alogia, have been a hallmark feature of schizophrenia, and are often associated with core negative symptoms and social impairment (7). Our results indicate that MFCCs capture information distinguishing depression, BD and schizophrenia from health.

The DB, DS, and BS models further evaluated the performances of voice features on pairwise classifications among the three mental disorders. DS (F-score = 0.83; AUC = 0.83) and BS (F-score = 0.91; AUC = 0.92) models showed promise. The BS model had the highest F-score and AUC across all seven models. However, the DB model performed the worst (F-score = 0.44), with an AUC score of 0.50 indicating voice features barely distinguished depression and BD. Further pairwise model comparisons showed no significant differences among the AH, DH, BH, SH, DS, and BS models. However, DB model performance was significantly worse than that of BS. Results indicate that voice features could help distinguish depression from disorders with similar mood symptoms.

Differing symptoms and vocal characteristics likely explain the results of differential diagnosis. For example, depression is associated with monotonous, hoarse, breathy or slurred speech reflecting anhedonia and sadness (5358). Schizophrenia is linked to poverty of speech, increased pauses, distinctive tone/intensity associated with core negative symptoms like flat affect, decreased emotional expression and difficulty controlling speech to express emotion properly (5965). Given that Espinola et al. (25) found vocal features distinguishing major depressive disorder, schizophrenia and other disorders, our MFCC derived i-vector approach may reveal subtle differences in anhedonic/sad depression symptoms versus alogia/flat affect in schizophrenia. However, further voice difference investigation between depression and schizophrenia is still needed to offer more explanation.

Another possible reason is that the i-vectors catch disorder-relevant information. To extract i-vectors, first the method learns shared case/control information, then removes shared components, yielding i-vectors capturing key individual differences (29, 43). Here, i-vectors captured distinct vocal information for different mental disorders.

Overall, vocal features could provide clinical value in distinguishing depression from selected disorders, and model improvement is helpful, especially for closely related conditions. This analysis establishes a foundation for future studies exploring vocal biomarkers for differential diagnosis of psychiatric disorders.

This research was unable to clearly differentiate between depression and BD. There appears to be substantial symptom overlap between these two conditions. Furthermore, the oscillation between manic and depressive states in BD can be subtle, as demonstrated by Grande et al. (66). Researchers have posited that there might be continued presence of subsyndromal residual symptoms during recovery from major affective episodes in bipolar disorder, and residual symptoms after resolution of a major affective episode indicate that the individual is at significant risk for a rapid relapse and/or recurrence, augmenting the intricacy of symptom presentation even during euthymia (67). Depressive symptoms may have been present in euthymic bipolar patients in this study, confounding diagnostic classification. This observation highlights the complexity of the euthymic bipolar condition. Future research should examine symptom and voice differences between BD phases and depression. A nuanced understanding of markers that distinguish unipolar depression from BD could sharpen diagnostic precision.

This study has limitations. We examined only depression, BD and schizophrenia in a small cultural sample, limiting generalizability. It is worth noting that this study lacks an anxiety disorder group, which also exhibits affective symptoms similar to depression. As the data came from a previous project in China, anxiety diagnoses were not included. Future work will gather systematic data on symptoms, clinical phase, psychological factors like emotion, cognition, and severity in these and other disorders, such as anxiety. This could yield insights into voice differences between depression and other conditions, enhancing the clinical value of vocal biomarkers.

Conclusion

This research systematically explored the ability of vocal biomarkers to distinguish depression from disorders with similar affective symptoms. Findings suggest vocal features could aid differential diagnosis for depression in clinical practice. Future research should investigate mechanisms by which vocal features differentiate depression and other disorders, and develop more advanced machine learning models so voice can enhance clinical diagnosis.

Funding

This work was supported by the Fundamental Research Funds for the Central Universities (CCNU21XJ021), Knowledge Innovation Program of Wuhan-Shuguang Project (2022020801020288), and the Research Program Funds of the Collaborative Innovation Center of Assessment toward Basic Education Quality (2022-04-030-BZPK01).

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Statements

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving human participants were reviewed and approved by the Institutional Review Board (IRB) of Institute of Psychology, Chinese Academy of Sciences. The patients/participants provided their written informed consent to participate in this study.

Author contributions

WP was in charge of overall research, data analysis, and drafting the paper. FD was in charge of offering professional guidance and advice for mental illnesses as a clinical specialist. XW helped on data analysis. BH and WZ helped with the analysis and revising the manuscript. TZ was in charge of collecting and organizing the data. All authors contributed to the article and approved the submitted version.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  • 1.

    BagbyRMRyderAGSchullerDRMarshallMB. The Hamilton depression rating scale: has the gold standard become a lead weight?Am J Psychiatr. (2004) 161:216377. doi: 10.1176/appi.ajp.161.12.2163

  • 2.

    FriedEI. The 52 symptoms of major depression: lack of content overlap among seven common depression scales. J Affect Disord. (2017) 208:1917. doi: 10.1016/j.jad.2016.10.019

  • 3.

    KendlerKSAggenSHFlintJBorsboomDFriedEI. The centrality of DSM and non-DSM depressive symptoms in Han Chinese women with major depression. J Affect Disord. (2018) 227:73944. doi: 10.1016/j.jad.2017.11.032

  • 4.

    KimYKParkSC. An alternative approach to future diagnostic standards for major depressive disorder. Prog Neuro-Psychopharmacol Biol Psychiatry. (2020) 2020:110133. doi: 10.1016/j.pnpbp.2020.110133

  • 5.

    MitchellAJVazeARaoS. Clinical diagnosis of depression in primary care: a meta-analysis. Lancet. (2009) 374:60919. doi: 10.1016/S0140-6736(09)60879-5

  • 6.

    LiuXFaesLKaleAUWagnerSKFuDJBruynseelsAet al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis. Lancet Digit Health. (2019) 1:e27197. doi: 10.1016/S2589-7500(19)30123-2

  • 7.

    ParolaASimonsenABlikstedVFusaroliR. Voice patterns in schizophrenia: a systematic review and Bayesian meta-analysis. Schizophr Res. (2020) 216:2440. doi: 10.1016/j.schres.2019.11.031

  • 8.

    CumminsNSchererSKrajewskiJSchniederSEppsJQuatieriTF. A review of depression and suicide risk assessment using speech analysis. Speech Comm. (2015) 71:1049. doi: 10.1016/j.specom.2015.03.004

  • 9.

    Faurholt-JepsenMBuskJFrostMVinbergMChristensenEMWintherOet al. Voice analysis as an objective state marker in bipolar disorder. Nat Publ Group. (2016) 6:856. doi: 10.1038/tp.2016.123

  • 10.

    MaxhuniAMuñoz-MeléndezAOsmaniVPerezHMayoraOMoralesEF. Classification of bipolar disorder episodes based on analysis of voice and motor activity of patients. Pervasive Mobile Comput. (2016) 4:841. doi: 10.1016/j.pmcj.2016.01.008

  • 11.

    RapcanVD’ArcySYeapSAfzalNThakoreJReillyRB. Acoustic and temporal analysis of speech: a potential biomarker for schizophrenia. Med Eng Phys. (2010) 32:10749. doi: 10.1016/j.medengphy.2010.07.013

  • 12.

    AfshanAGuoJParkSJRaviVFlintJAlwanA. Effectiveness of voice quality features in detecting depression. Interspeech. (2018) 2018:167680 doi: 10.21437/Interspeech.2018-1399

  • 13.

    HeLCaoC. Automated depression analysis using convolutional neural networks from speech. J Biomed Inform. (2018) 83:10311. doi: 10.1016/j.jbi.2018.05.007

  • 14.

    JiangHHuBLiuZWangGZhangLLiXet al. Detecting depression using an ensemble logistic regression model based on multiple speech features. Comput Math Methods Med. (2018) 2018:6508319. doi: 10.1155/2018/6508319

  • 15.

    PanWFlintJShenhavLLiuTLiuMHuBet al. Re-examining the robustness of voice features in predicting depression: compared with baseline of confounders. PLoS One. (2019) 14:e0218172. doi: 10.1371/journal.pone.0218172

  • 16.

    RohaniDAFaurholt-JepsenMKessingLVBardramJE. Correlations between objective behavioral features collected from mobile and wearable devices and depressive mood symptoms in patients with affective disorders: systematic review. JMIR Mhealth Uhealth. (2018) 6:9691. doi: 10.2196/mhealth.9691

  • 17.

    TaguchiTTachikawaHNemotoKSuzukiMNaganoTTachibanaRet al. Major depressive disorder discrimination using vocal acoustic features. J Affect Disord. (2018) 225:21420. doi: 10.1016/j.jad.2017.08.038

  • 18.

    HuangYWangYUWangHLiuZYuXYanJet al. Prevalence of mental disorders in China: a cross-sectional epidemiological study. Lancet Psychiatry. (2019) 6:21124. doi: 10.1016/S2215-0366(18)30511-X

  • 19.

    El AyadiMKamelMSKarrayF. Survey on speech emotion recognition: features, classification schemes, and databases. Pattern Recogn. (2011) 44:57287. doi: 10.1016/j.patcog.2010.09.020

  • 20.

    SwainMRoutrayAKabisatpathyP. Databases, features and classifiers for speech emotion recognition: a review. Int J Speech Technol. (2018) 21:93120. doi: 10.1007/s10772-018-9491-z

  • 21.

    American Psychiatric Association. Diagnostic and statistical manual of mental disorders (DSM-IV) (1994). Washington, D.C: American Psychiatric Pub

  • 22.

    EspinolaCWGomesJCPereiraJMSdos SantosWP. Vocal acoustic analysis and machine learning for the identification of schizophrenia. Res Biomed Eng. (2021) 37:3346. doi: 10.1007/s42600-020-00097-1

  • 23.

    WeinerLGuidiADoignon-CamusNGierschABertschyGVanelloN. Vocal features obtained through automated methods in verbal fluency tasks can aid the identification of mixed episodes in bipolar disorder. Transl Psychiatry. (2021) 11:415. doi: 10.1038/s41398-021-01535-z

  • 24.

    HiguchiMTokunoSHINICHINakamuraMShinoharaSHUJIMitsuyoshiSOmiyaYet al. Classification of bipolar disorder, major depressive disorder, and healthy state using voice. Asian J Pharm Clin Res. (2018) 11:8993. doi: 10.22159/ajpcr.2018.v11s3.30042

  • 25.

    Wanderley EspinolaCGomesJCMônica Silva PereiraJdos SantosWP. Detection of major depressive disorder, bipolar disorder, schizophrenia and generalized anxiety disorder using vocal acoustic analysis and machine learning: an exploratory study. Res Biomed Eng. (2022) 38:81329. doi: 10.1007/s42600-022-00222-2

  • 26.

    ZhuYKimYCProctorMINarayananSSNayakKS. Dynamic 3D visualization of vocal tract shaping during speech. IEEE Trans Med Imaging. (2013) 32:83848. doi: 10.1109/TMI.2012.2230017

  • 27.

    TanhaJAbdiYSamadiNRazzaghiNAsadpourM. Boosting methods for multi-class imbalanced data classification: an experimental review. J Big Data. (2020) 7:147. doi: 10.1186/s40537-020-00349-y

  • 28.

    BatistaGEPratiRCMonardMC. A study of the behavior of several methods for balancing machine learning training data. ACM SIGKDD Explor Newsletter. (2004) 6:209. doi: 10.1145/1007730.1007735

  • 29.

    DiYWangJLiWZhuT. Using i-vectors from voice features to identify major depressive disorder. J Affect Disord. (2021) 288:1616. doi: 10.1016/j.jad.2021.04.004

  • 30.

    CumminsN.EppsJ.BreakspearM.GoeckeR. (2011). An investigation of depressed speech detection: features and normalization. In 12th Annual Conference of the International Speech Communication Association

  • 31.

    CumminsN.EppsJ.SethuV.KrajewskiJ. (2014). Variability compensation in small data: oversampled extraction of i-vectors for the classification of depressed speech. In 2014 IEEE international conference on acoustics, speech and signal processing (ICASSP) (970974). IEEE.

  • 32.

    JoshiJGoeckeRAlghowinemSDhallAWagnerMEppsJet al. Multimodal assistive technologies for depression diagnosis and monitoring. J Multimodal User Interf. (2013) 7:21728. doi: 10.1007/s12193-013-0123-2

  • 33.

    NasirMJatiAShivakumarPGChakravarthulaSNGeorgiouP. Multimodal and multiresolution depression detection from speech and facial landmark features. AVEC 2016 – proceedings of the 6th international workshop on audio/visual emotion challenge. Co-Located ACM Multimedia. (2016) 2016:4350. doi: 10.1145/2988257.2988261

  • 34.

    HashimN.WahidahN. N. (2017). Analysis of power spectrum density of male speech as indicators for high risk and depressed decision (doctoral dissertation). Vanderbilt University.

  • 35.

    LiuZ.HuB.LiX.LiuF.WangG.YangJ. (2017). Detecting depression in speech under different speaking styles and emotional valences. International conference on brain informatics (261271). Springer, Cham.

  • 36.

    SimantirakiO.CharonyktakisP.PampouchidouA.TsiknakisM.CookeM. (2017). Glottal source features for automatic speech-based depression assessment. INTERSPEECH (27002704). Stockholm, Sweden.

  • 37.

    PanWWangJLiuTLiuXLiuMHuBet al. Depression recognition based on speech analysis. Chin Sci Bull. (2018) 63:208192. doi: 10.1360/N972017-01250

  • 38.

    WangJZhangLLiuTPanWHuBZhuT. Acoustic differences between healthy and depressed people: a cross-situation study. BMC Psychiatry. (2019) 19:300. doi: 10.1186/s12888-019-2300-7

  • 39.

    GoodPI. Permutation, parametric and bootstrap tests of hypotheses: a practical guide to resampling methods for testing hypotheses (2005) 100

  • 40.

    CenzerIBoscardinWJBergerK. Performance of matching methods in studies of rare diseases: a simulation study. Intractable & Rare Diseases. Research. (2020) 9:7988. doi: 10.5582/irdr.2020.01016

  • 41.

    RStudioT. RStudio: Integrated development for R. Boston, MA: Rstudio Team, PBC (2020) Available at:https://www.rstudio.com.

  • 42.

    MatlabS. Matlab. Natick, MA: The MathWorks (2012).

  • 43.

    DehakNKennyPJDehakRDumouchelPOuelletP. Front-end factor analysis for speaker verification. IEEE Trans Audio Speech Lang Process. (2011) 19:78898. doi: 10.1109/TASL.2010.2064307

  • 44.

    PoveyD.GhoshalA.BoulianneG.BurgetL.GlembekO.GoelN.et al, (2011). The Kaldi speech recognition toolkit. IEEE 2011 workshop on automatic speech recognition and understanding (no. CONF). IEEE Signal Processing Society.

  • 45.

    FriedmanJHastieTTibshiraniR. Regularization paths for generalized linear models via coordinate descent. J Stat Softw. (2010) 33:122. PMID:

  • 46.

    DehejiaRHWahbaS. Propensity score-matching methods for nonexperimental causal studies. Rev Econ Stat. (2002) 84:15161. doi: 10.1162/003465302317331982

  • 47.

    AlghowinemS.GoeckeR.WagnerM.EppsJ.BreakspearM.ParkerG. (2013). Detecting depression: a comparison between spontaneous and read speech. ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing – Proceedings, 75477551

  • 48.

    HorwitzR.QuatieriT. F.HelferB. S.YuB.WilliamsonJ. R.MundtJ. (2013). On the relative importance of vocal source, system, and prosody in human depression. 2013 IEEE international conference on body sensor networks, BSN 2013

  • 49.

    QuatieriT. F.MalyskaN. (2012). Vocal-source biomarkers for depression: A link to psychomotor activity. In 13th Annual Conference of the International Speech Communication Association.

  • 50.

    SidorovMMinkerW. Emotion recognition and depression diagnosis by acoustic and visual features: a multimodal approach. AVEC 2014 – proceedings of the 4th international workshop on audio/visual emotion challenge. Workshop MM. (2014) 2014:816. doi: 10.1145/2661806.2661816

  • 51.

    Faurholt-JepsenMRohaniDABuskJVinbergMBardramJEKessingLV. Voice analyses using smartphone-based data in patients with bipolar disorder, unaffected relatives and healthy control individuals, and during different affective states. Int J Bipolar Disorders. (2021) 9:113. doi: 10.1186/s40345-021-00243-3

  • 52.

    TahirYYangZChakrabortyDThalmannNThalmannDManiamYet al. Non-verbal speech cues as objective measures for negative symptoms in patients with schizophrenia. PLoS One. (2019) 14:e0214314. doi: 10.1371/journal.pone.0214314

  • 53.

    FranceDJShiaviRG. Acoustical properties of speech as indicators of depression and suicidal risk. IEEE Trans Biomed Eng. (2000) 47:82937. doi: 10.1109/10.846676

  • 54.

    LowL. S. A.MaddageN. C.LechM.SheeberL.AllenN. (2010). Influence of acoustic low-level descriptors in the detection of clinical depression in adolescents. ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings, 51545157

  • 55.

    MooreE.ClementsM.PeiferJ.WeisserL. (2003). Analysis of prosodic variation in speech for clinical depression. Annual International Conference of the IEEE Engineering in Medicine and Biology – Proceedings, 3, 29252928

  • 56.

    MundtJCSnyderPJCannizzaroMSChappieKGeraltsDS. Voice acoustic measures of depression severity and treatment response collected via interactive voice response (IVR) technology. J Neurolinguistics. (2007) 20:5064. doi: 10.1016/j.jneuroling.2006.04.001

  • 57.

    OzdasAShiaviRGSilvermanSESilvermanMKWilkesDM. Investigation of vocal jitter and glottal flow spectrum as possible cues for depression and near-term suicidal risk. IEEE Trans Biomed Eng. (2004) 51:153040. doi: 10.1109/TBME.2004.827544

  • 58.

    TrevinoACQuatieriTFMalyskaN. Phonologically-based biomarkers for major depressive disorder. EURASIP J Adv Signal Process. (2011) 2011:42. doi: 10.1186/1687-6180-2011-42

  • 59.

    AlpertMRosenbergSDPougetERShawRJ. Prosody and lexical accuracy in flat affect schizophrenia. Psychiatry Res. (2000) 97:10718. doi: 10.1016/S0165-1781(00)00231-6

  • 60.

    CohenASMitchellKRDochertyNMHoranWP. Vocal expression in schizophrenia: less than meets the ear. J Abnorm Psychol. (2016) 125:299309. doi: 10.1037/abn0000136

  • 61.

    CohenASNajoliaGMKimYDinzeoTJ. On the boundaries of blunt affect/alogia across severe mental illness: implications for research domain criteria. Schizophr Res. (2012) 140:415. doi: 10.1016/j.schres.2012.07.001

  • 62.

    GalynkerIICohenLJCaiJ. Negative symptoms in patients with major depressive disorder: a preliminary report. Neuropsychiatry Neuropsychol Behav Neurol. (2000) 13:1716. PMID: Available at: https://pubmed.ncbi.nlm.nih.gov/10910087/

  • 63.

    HoekertMKahnRSPijnenborgMAlemanA. Impaired recognition and expression of emotional prosody in schizophrenia: review and meta-analysis. Schizophr Res. (2007) 96:13545. doi: 10.1016/j.schres.2007.07.023

  • 64.

    MillanMJFoneKStecklerTHoranWP. Negative symptoms of schizophrenia: clinical characteristics, pathophysiological substrates, experimental models and prospects for improved treatment. Eur Neuropsychopharmacol. (2014) 24:64592. doi: 10.1016/j.euroneuro.2014.03.008

  • 65.

    TrémeauFMalaspinaDDuvalFCorrêaHHager-BudnyMCoin-BariouLet al. Facial expressiveness in patients with schizophrenia compared to depressed patients and nonpatient comparison subjects. Am J Psychiatr. (2005) 162:92101. doi: 10.1176/appi.ajp.162.1.92

  • 66.

    GrandeIBerkMBirmaherBVietaE. Bipolar disorder. Lancet. (2016) 387:156172. doi: 10.1016/S0140-6736(15)00241-X

  • 67.

    JuddLLSchettlerPJAkiskalHSCoryellWLeonACMaserJDet al. Residual symptom recovery from major affective episodes in bipolar disorders and rapid episode relapse/recurrence. Arch Gen Psychiatry. (2008) 65:38694. doi: 10.1001/archpsyc.65.4.386

Summary

Keywords

depression, healthy controls, schizophrenia, bipolar disorder, i-vectors, logistic regression MFCCs

Citation

Pan W, Deng F, Wang X, Hang B, Zhou W and Zhu T (2023) Exploring the ability of vocal biomarkers in distinguishing depression from bipolar disorder, schizophrenia, and healthy controls. Front. Psychiatry 14:1079448. doi: 10.3389/fpsyt.2023.1079448

Received

25 October 2022

Accepted

30 June 2023

Published

20 July 2023

Volume

14 - 2023

Edited by

Ronghui Liu, University of Chinese Academy of Sciences, China

Reviewed by

Paul Thuras, United States Department of Veterans Affairs, United States; Ang Li, Beijing Forestry University, China

Updates

Copyright

*Correspondence: Tingshao Zhu,

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics