Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Psychiatry, 12 August 2025

Sec. Psychopathology

Volume 16 - 2025 | https://doi.org/10.3389/fpsyt.2025.1633492

Emotional speech markers of psychiatric disturbance in Huntington’s disease

Lucie Chenain,,,,*Lucie Chenain1,2,3,4,5*Audrey Fabre,,,,Audrey Fabre1,2,4,6,7Hadrien Titeux,,,,Hadrien Titeux1,2,4,6,8Graa MorgadoGraça Morgado7Katia Youssov,,,,Katia Youssov1,2,4,6,7Chlo ClavelChloé Clavel3Anne-Catherine Bachoud-Lvi,,,,Anne-Catherine Bachoud-Lévi1,2,4,6,7
  • 1NeuroPsychologie Interventionnelle, Département d’Etudes Cognitives, Ecole Normale Supérieure, Paris Sciences et Lettres University, Paris, France
  • 2NeuroPsychologie Interventionnelle, Université Paris Est Créteil, Institut national de la santé et de la recherche médicale (INSERM) Unité 955, Institut Mondor de Recherche Biomédicale, Créteil, France
  • 3Automatic Language Modelling and Analysis & Computational Humanities (ALMAnaCH), Institut national de recherche en sciences et technologies du numérique (INRIA), Paris, France
  • 4NeurATRIS, Hôpital Henri Mondor, Créteil, France
  • 5Learning Planet Institute, Université Paris Cité, Paris, France
  • 6Assistance Publique — Hôpitaux de Paris (AP-HP), Hôpital Henri Mondor-Albert Chenevier, Centre de Référence Maladie de Huntington, Service de Neurologie, Créteil, France
  • 7Inserm, Centre d’Investigation Clinique 1430, AP-HP, Hôpital Henri Mondor, Créteil, France
  • 8Laboratoire de Sciences Cognitives et Psycholinguistique, Centre national de la recherche scientifique (CNRS), Unité 8554, Paris Sciences et Lettres (PSL) University, Paris, France

Introduction: Psychiatric disorders and difficulties in emotional expression represent a major problem in the management of Huntington’s Disease (HD). To improve patient follow-up, we propose to investigate the link between emotional expression and psychiatric symptoms, measured by the Problem Behaviors Assessment (PBA) scale. To this aim we developed the first emotional/psychiatric speech corpus, emoHD.

Methods: We included 102 HD gene carriers and 35 healthy controls (HC). Psychiatric symptoms were assessed using PBA sub-scales for Depression, Irritability/aggressivity, Apathy, and Obsessive/compulsive symptoms. Speech was annotated using three emotional descriptors: primary emotions, affective phenomena, and activation levels. Affective phenomena labels were selected based on PBA statements by external participants unaware of the study’s aims. We analyzed (1) emotional descriptors’ relationships, (2) emotional expression differences between HD and HC, and (3) the associations between emotions and psychiatric symptoms.

Results: HD patients showed reduced emotional expressiveness than HC with more neutral activation levels (=0). Only the primary emotion “angry” was less expressed in HD compared to HC. In contrast they expressed more affective phenomena states like apathetic, confused, “depressed”, “disoriented”, “frustrated”, and “pessimistic” than HC, whereas they expressed less “other” and “irritable” than HC. Expressed emotions were congruent with psychiatric symptoms (e.g., “anxious” and “nervous” are positively associated with Depression PBA sub-scale; “frustrated” with Irritability/aggressivity PBA sub-scale).

Conclusion: We showed that speech is a promising marker for emotional/psychiatric symptoms in HD, supporting future remote monitoring and personalized care strategies.

1 Introduction

Huntington’s Disease (HD) is an autosomal dominant neurodegenerative disorder caused by the repeated expansion of CAG trinucleotides in the huntingtin gene, with full penetrance achieved with ≧ 39 repetitions (1). It is characterized by motor, cognitive and psychiatric symptoms, leading to disability and gene carriers’ death close to 20 years after disease onset (2). Its large spectrum of symptoms makes it a valuable model for studying other neurodegenerative disorders.

Indeed, HD individuals experience both psychiatric disorders and emotional production and perception deficits for facial, body, and vocal expressions (37). On one hand, ineffective emotional processing can impact interpersonal relations and contribute to emotional reasoning difficulties, which may underlie or interact with patients’ psychiatric profiles (8). On the other, HD individuals exhibit risky behaviors: they have the highest rate of suicide among those with a diagnosed neurological disorder, with an absolute risk of 1.6 (9). Irritability is common, affecting 38-73% of individuals, and may lead to hetero-aggressive and auto-aggressive behaviors. Other psychiatric behaviors include depression (prevalence of 20-56%), anxiety (16.7-24%), apathy (34-76%), and psychosis (10.4%) (8). These symptoms affect patients’ daily functioning (1012), quality of life (1315), and relationships, often causing social withdrawal, and family breakups (1618). While pharmaceutical treatments and psychological support may relieve symptoms (19), their detection and monitoring remains a challenge, presumably, because of patients’ difficulties in expressing emotion. Yet, the absence of link between emotions and psychiatric behaviors represents a limitation.

This highlights the need to investigate the interplay between emotional deficits and psychiatric manifestations in HD. Several factors make speech a valuable medium for gathering such information. Speech carries a communicative message through linguistic content along with paralinguistic cues, which are defined as nonverbal forms of communication that contribute to conveying the message’s meaning. Its production is intrinsically influenced by a speaker’s physical and psychological status changes, altering their vocal apparatus control, even subconsciously (20). Speech recordings of individuals with self-reported or clinically diagnosed psychiatric disorders have been used to analyze psychiatric symptoms. This has contributed to the development of speech corpora with non-dysarthric patients, such as the AVEC, RAPID and DEPAC corpora for depression (2123), MONARCA for bipolar disorder (24), Etude 1000 (25) for post-traumatic stress disorder, and the BBRS (26) or Scherer et al. (27) for suicide. Such databases can identify psychiatric behaviors (28, 29), support differential diagnosis (30, 31) and predict the onset of a psychiatric disorder in at-risk individuals (31). Speech also carries acoustic cues of emotions, such as prosody, intensity, and speech rate (32), motivating the development of emotional speech corpora where recordings are annotated with emotional descriptors to capture various emotional information (33). In neurological populations, such corpora are used for emotional analysis (34) and the development of speech emotion recognition models (35, 36). However, the underrepresentation of pathological speakers and the lack of psychiatric information collected (3741) can make generalization and real-world applications difficult. The development of disorder-specific emotional speech corpora that include clinical and psychiatric assessment of speakers could therefore improve the generalizability and applicability of findings.

We developed the HD emotional speech corpus (emoHD) using recordings from participants psychiatrically evaluated by certified neurologists, with emotional annotations provided by speech therapists blinded to the neurological assessment. Emotions were annotated using three emotional descriptors: first, primary emotions were annotated, referring to universal, cross-cultural categories such as Ekman’s six (42), providing a basic emotional characterization. Second, affective phenomena, as defined by Scherer (43), which include a broader range of emotional and affective states including: feelings referring to subjective experience of emotion; moods which are diffuse and longer lasting affective states; preferences which are stable evaluative judgments; attitudes for long-lasting predispositions or beliefs; affective dispositions which are personality traits predisposing individuals to certain emotions or moods; and interpersonal stances which are affective styles that develop during interactions with others. To select a list of affective phenomena labels, external participants reviewed statements extracted from the standardized HD psychiatric scale (the Problem Behaviors Assessment - short form (PBA) (44)) and identified the emotional manifestations they perceived in each. These responses informed the selection of labels used to annotate patients’ speech. These external participants were blind to the study’s objectives, unfamiliar with HD, and did not know the clinical significance of the PBA. Finally, speech therapists rated the activation levels, representing the emotional excitation expressed through physiological changes (heart rate increase, internal temperature rising, etc.) (45) (Figure 1).

Figure 1
Flowchart illustrating data collection process. Section A describes affective phenomena selection involving PBA statements and external participants proposing labels. Section B involves emotional annotations by neuropsychologists and speech therapists, categorizing speech into primary emotions, affective phenomena, and activation levels. Section C covers psychiatric evaluation by neurologists, assessing conditions like depression and apathy. Participants, labeled HD and HC, provide data. The study explores associations between emotions and psychiatric symptoms, with databases for speech and clinical information.

Figure 1. Overview of the emoHD corpus. Participants in the emoHD corpus were Huntington’s Disease (HD) gene carriers’ and healthy controls (HC). They underwent psychiatric evaluation (pathway C) by certified neurologists (orange box) using the Problem Behaviors Assessment (PBA) sub-scales, constituting the clinical database of emoHD. Independently, participants’ speech was recorded by neuropsychologists (green box), constituting the speech database of emoHD. Recordings were emotionally annotated (pathway B) by speech therapists (green box) which included three emotional descriptors: (1) primary emotions for coarse-grain characterization; affective phenomena to capture emotional manifestations specific to Huntington’s disease — these labels were selected by external participants (grey box) to maintain all steps independent and blind to each other (pathway A); and (3) activation levels to indicate the emotional excitation expressed. Researcher (LC) was blind to the participants’ status, and ran the statistical analysis.

With emoHD, we (1.) investigated and validated the annotation scheme by establishing how the different emotional descriptors are associated with each other; (2.) explored how emotional expression differs between HD gene carriers and controls; and (3.) assessed how emotional expression relates to psychiatric profiles. This allowed us to propose a framework for the use of emotional speech analysis as a potential marker for psychiatric symptoms in HD to constitute the first step towards patient remote monitoring of emotional and affective disturbances based on speech analysis. It will likely contribute to targeted interventions and personalize care strategies for individuals living with HD and may ultimately be transferred to other neurodegenerative diseases.

2 Materials and methods

2.1 Data collection

2.1.1 Participants

The emoHD cohort collected clinical information and recordings from 137 participants from two observational cohorts REPAIR-HD and BIO-HD (NCT03119246 & NCT01412125) at the Henri Mondor Hospital (Créteil, France), with optional longitudinal follow-up. To constitute the EmoHD cohort, we included only the first visits of subsequent participants who had performed on the same day both speech recordings and a general clinical evaluation, using the Unified Huntington’s Disease Rating Scale (UHDRS) (46), the international reference scale in HD. All participants were French native speakers, aged ≥ 18 years, and signed an informative consent form. Exclusion criteria were incapacity to consent and significant neurological or psychiatric comorbidities unrelated to HD. Ethics approval was given by the institutional review board from Henri Mondor Hospital (Créteil, France) for the BIO-HD cohort and the CPP Saint Louis French part of the REPAIR-HD cohort, in compliance with the Helsinki Declaration.

Participants consisted of 102 HD gene carriers at various stages with a CAG repeat expansion ≥ 40 and 35 healthy controls (HC) without any cognitive deficits [Mattis Dementia Rating Scale ≥ 136 (47)]. HD gene carriers were classified on the Integrated Staging System (HD ISS): 23 gene carriers were classified in stages 0–1 corresponding to no symptoms or biomarker of cerebral volume pathogenesis, 25 in stage 2 showing symptoms of HD and 54 in stage 3 with functional change associated to HD (respectively HD ISS 0-1, 2, and 3) (48).

In addition, 96 external participants were recruited to answer an online Google form to establish a list of affective phenomena related to HD (see 2.2). They were French-speaking participants unfamiliar with HD and the PBA, blind to the study’s objective, and not working in a health-related field (see Supplementary Table 1).

2.1.2 General clinical evaluation

Certified neurologists evaluated motor symptoms with the UHDRS Total Motor Score (TMS) and daily life functioning with the Total Functional Capacity (TFC) (46, 49). Cognitive evaluation was performed with the UHDRS verbal fluency (under 1 minute), the Stroop word, color, interference test, and the symbol digit modalities test (SDMT) (46).

Psychiatric evaluation was conducted using the PBA (44), a semi-structured diagnostic questionnaire designed to assess eleven behavioral items specific to HD: depressive mood, suicidal ideation, anxiety, irritability, angry or aggressive behavior, apathy, perseverative thinking or behavior, obsessive-compulsive behavior, delusions, hallucinations, and disoriented behavior. Neurologists rated the frequency of each behavioral item (0 = never/almost never, 1= seldom (less than 1/week), 2= sometimes (up to 4/week), 3 = frequently (most days or 6 to 7/week), 4 = daily or almost daily) and its severity (0 = absent, 1 = slight, questionable, 2 = mile (present, not a problem), 3 = moderate (symptom causing problem), 4 = severe (almost intolerable for carer).

The behavioral items scores are combined to create five PBA sub-scale scores, which are calculated as the sum of the product of each item’s severity and frequency (50, 51): the Depression sub-scale combining depressive mood, suicidal ideation and anxiety; the Irritability/aggressivity sub-scale gathering irritability and angry or aggressive behavior items, the Apathy sub-scale with the apathy item only; and lastly the Obsessive/compulsive sub-scale with perseverative thinking or behavior and obsessive-compulsive behaviors. We excluded the psychosis sub-scale, with delusions and hallucinations, as only one participant scored ≧ 1.

2.1.3 Speech recordings

The speech recordings were collected as part of the BasalVoice speech recording protocol (52, 53), which included 8 speech tasks (approximately 10 minutes) among which 5 were retained for the emoHD corpus. The order of the tasks altered elicited emotional and non-emotional speech to avoid emotional contamination between tasks: counting from 1–20 forward/backwards, narrating one’s last 24 hours, then a sad story, reciting the twelve months in a year forward/backwards with hands in front and eyes closed, narrating the little riding hood tale, an anger story, describing the cookie theft picture (CTP) (54), and finally narrating a happy story to avoid ending the interview with a negative emotion. Three tasks were not considered in this study: the little riding hood task whose assessment varied among participants depending on whether they used a picture for assistance, and the forward/backward counting and month recitation which were designed to increase cognitive load by requiring participants to lift their hands and close their eyes.

Emotional speech tasks consisted of narrative prompts based on Ekman’s set of basic emotions. The narration of emotions is an effective way for speakers to express, understand and share them. The successful completion of that task favors the elicitation of emotions in speakers (55, 56). In other terms, if someone really aims to tell a sad story, we expect that they should express markers of sadness at some points. Instructions were for the sad story: “Can you tell me a story that you find sad, really unhappy, or really depressing?”; the anger story: “Can you tell me a story that you find irritating, upsetting, a really annoying story?; and the happy story: “Can you tell me a pleasant story, full of happiness and joy, something uplifting?”. Relevance of eliciting fear, disgust, and surprise did not appear relevant regarding the behavioral evaluation.

Non-emotional speech tasks included eliciting speech without emotional prompting by having individuals narrate their last 24 hours (“Can you tell me what you’ve been doing for the last 24 hours?”). This recorded natural expressions of neutral or emotional speech, depending on their recent personal experience. The CTP from the Boston Diagnostic Aphasia Examination traditionally assesses language abilities in aphasia (54) and is widely used as a speech evaluation in neurodegenerative diseases (57). Participants were shown the CTP picture and instructed “Could you please describe everything you see in the picture?”.

All participants were recorded under the same conditions in a hospital consultation room just after their neuropsychological assessment. They sat at a table facing the neuropsychologist, with the recording device placed 20 cm from the participant. The device used was a ZOOM H4n Pro recorder, with a sampling rate of 44.1 kHz and a 16-bit resolution. The neuropsychologists could easily set up the recording device without the need of an acoustic engineer. Participants could stop the session anytime. The duration of each speech task in recordings was extracted with python SpontaneousCHAT package.

2.2 Annotation protocol

A preliminary transcription of the linguistic content was performed by speech therapists using Praat (58) following the BasalVoice protocol (52, 53). The emoHD annotation campaign was managed on Seshat (59), a tool to automate the management of annotation campaigns of audio and speech data, by managing file handling, non-conformity and annotation inconsistency. Three speech therapists, blinded to the study’s aim, listened to the recordings and divided them in emotional segments on Praat, defined as sections of speech with a homologous emotional content (60). They were asked to label each segment with the three emotional descriptors:

- First, during the initial listening and segmentation, the speech therapists labeled each segment in one of the emotions intended to be elicited by the emotional speech tasks, namely triste (sad), joie (happy), colère (angry), adding neutre (neutral) to infer a referential neutral emotional state and autre (other) for all other emotions.

- Then, during a second listening of the recordings, the speech therapists annotated the same emotional segments with one or more labels corresponding to affective phenomena. The list of affective phenomena labels was defined by external participants who were blind to the study. They provided the word describing the emotional manifestations for each of the 33 PBA statements, which corresponded to the severity scale, across the 11 PBA items, ranging from mild (=2) to severe (=4). (as described in 2.1.2). The order of appearance of the statements was randomized so that the participants remained blind to the item and its severity. After converting all answers in their adjective form, the affective phenomena labels with a frequency ≧ 10% per statement were kept. This yielded a list of 23 affective phenomena labels related to HD behavior, which were used for annotation: triste (sad), déprimé.e (depressed), désespéré.e (hopeless), pessimiste (pessimistic), détresse (distressed), anxieux.se (anxious), inquiet (worried), stressé.e (stressed), angoissé.e (nervous), peureux.se (fearful), irritable (irritable), coléreux.se (quick tempered), frustré.e (frustrated), rageant.e (infuriating), violent.e (violent), apathique (apathetic), obsessionnel.le (obsessive), délirant (delusional), désorienté.e (disoriented), perdu.e (lost), and confus.e (confused). We also added the affective phenomena autre (other affective phenomena) and neutre (neutral affective phenomena) as complementary.

- Last, activation levels of each segment were scored from 0 (no activation, equivalent to neutral) to 3 for high activation, representing the emotional excitation expressed through physiological changes (45).

The speech therapists could also add comments to the emotional segments which could be other emotions or observations.

This approach resulted in an unbiased selection of labels with intelligible common vocabulary. After annotating the whole corpus, the duration (in seconds) and the count of occurrence of each primary emotion, affective phenomenon and activation level were measured for each participant. To account for variations in recording length between individuals, the proportion of each emotional label was calculated by dividing the total duration of each label in the recording by the overall recording time.

The annotation protocol was pretested on 10 randomly selected recordings from participants’ follow-up visits (which were not included in the emoHD corpus). Agreement between the three speech therapists was obtained with the γ-agreement (61) using its pygamma python implementation (62) (Table 1). The γ-gamma is an agreement measure evaluating both the recording’s segmentation into emotional segments and their labeling, allowing for penalties in cases of discrepancies between annotators. In contrast to Cohen’s kappa or Krippendorff alpha, the γ-gamma has not been benchmarked yet, hence no threshold measure of agreement has been defined. Considering that the γ-gamma is more conservative than Cohen’s kappa or Krippendorff’s alpha, obtaining γ-gamma measures all equal or greater than 0.76 considered good.

Table 1
www.frontiersin.org

Table 1. γ-gamma inter-rater agreement.

2.3 Statistical analysis

We employed non-parametric statistical approaches to perform categorical associations, adapted to the sample size and characteristics of our population sample. This enhanced the statistical power of our analysis and accounted for the lack of distributional assumptions. We applied permutation testing to investigate the associations between emotional descriptors and emotional expression differences across HD clinical stages (63, 64).

2.3.1 Associations between emotional descriptors

We examined the correspondence between the labelling of emotional descriptors (primary emotions, affective phenomena and activation levels) in each emotional segment. We first conducted a two-sided Fisher’s exact test with simulated p-values with 10000 replications. We then ran a pairwise Fisher’s exact test to evaluate the individual associations between each pair of labels: primary emotion vs affective phenomena, affective phenomena vs activation and primary emotion vs activation. Each pairwise test was corrected for multiple comparisons, adjusting the p-values using Bonferroni correction to account for possible type I errors. Odds-Ratio (OR) of each comparison was estimated to facilitate the interpretation (65, 66).

2.3.2 Emotional expression in Huntington’s disease gene carriers and healthy controls

We explored the difference in emotional expressions between HD gene carriers and HC by conducting a 10000 permutation-based t-test on the proportions of emotional labels (primary emotions, affective phenomena, and activation levels). This approach randomly reshuffles group labels to generate a null distribution of t-statistics, allowing for the assessment of group differences without assuming normality. For the emotional labels with significant proportions differences (permutation p-values ≦ 0.05), we further proceeded with a 10000 permutation ANOVA comparing the HD ISS stages (HD ISS 0-1, 2 and 3) and controls, followed by pairwise comparison using Tukey’s test. This identified differences between specific clinical group status.

2.3.3 Association between emotional expression and psychiatric behaviors

We explored the association between emotional expression in the emoHD corpus and psychiatric symptoms with a multilinear regression analysis. The regressions were performed to test the associations between the proportions of primary emotions, affective phenomena and activation levels with the participants’ psychiatric sub-scale scores (Depression, Irritability/aggressivity, Apathy and Obsessive/compulsive).

For each psychiatric sub-scale, a multilinear regression used individuals’ proportions as predictors, while adjusting for sex, age and the clinical group status (HD ISS 0-1, 2, 3 and HC). A bidirectional stepwise selection method was applied to identify the best predictors explaining the variance (R2) in the sub-scale scores. The statistical significance of each predictor was defined with a p-value ≦ 0.05.

All statistical analysis were performed in R (version 4.4.2), using the stats, dplyr, multcomp, lmPerm, coin and ggplot2 and sjPlot packages. All preprocessing and statistical analysis are available on https://gitlab.cognitive-ml.fr/lchenain/emohd_corpus.git and are suited for python and R integrated development environments.

3 Results

3.1 Clinical assessment

The demographic and clinical information of the participants are summarized in Table 2. An ANOVA revealed significant differences in sex and age, the last explained by HD carriers in stages 0–1 being younger than more advanced HD stages and controls. Higher TMS and lower TFC scores, indicating worse motor and functional performance in HD carriers at stages 2 and 3, resulting in overall group differences. Similarly, cognitive evaluation indicated worse performances of HD carriers at stages 2 and 3 compared to HC and HD carriers in stage 0-1.

Table 2
www.frontiersin.org

Table 2. Demographics.

Among the 137 participants, 131 were assessed for psychiatric symptoms with the PBA. Table 3. describes the PBA sub-scales scores across groups. Higher scores were observed for the Depression, Apathy, and Obsessive/compulsive sub-scales as the disease stage advanced.

Table 3
www.frontiersin.org

Table 3. Problem Behaviors Assessment sub-scale scores.

3.2 Descriptive content of the emoHD corpus

The duration in seconds of the participants’ speech recordings in the emoHD corpus, excluding periods when the neuropsychologist was providing instructions or speaking, is shown in Table 4. The mean duration was similar across groups for the total recording duration, whereas the speech rate was lower for HD participants in the ISS 2 and 3 clinical stages compared to HC; and compared to HD ISS 0-1. Duration of the speech tasks can be found in the Supplementary Table 2.

Table 4
www.frontiersin.org

Table 4. Speech recordings in seconds.

Annotating emotions took approximately 30 minutes per participant’s recordings. Once annotated, the emoHD corpus contained 6965 emotional segments defined by the speech therapists, which were labeled for primary emotions, affective phenomena and activation levels. The average emotional segment was 5.79 seconds, with a standard deviation of 6.17 and a range varying from 0.13 to 68.01 seconds. Figure 2 displays the distribution (in seconds) of each possible label per speech tasks included in the emoHD corpus. The annotation of affective phenomena violent, obsessive and lost never occurred in the corpus.

Figure 2
Four bar charts display durations of emotions in seconds. Chart A shows primary emotions' durations, highlighting “Neutral” as the longest. Chart B presents activation levels, with level zero being the longest. Chart C details affective phenomena durations, with “Sad affective phenomena” being dominant. The fourth chart shows “Other” and “Neutral” affective phenomena, with “Neutral” lasting longer. Color-coded tasks represent different stories across all charts.

Figure 2. Distribution in seconds of primary emotions, affective phenomena and activation levels across the speech tasks. Distribution of primary emotions (A) and activation levels (B) across speech tasks. The distribution of affective phenomena is displayed separately for scaling reasons: neutral and other affective phenomena [Panel (C) right] and all other affective phenomena [Panel (C) left].

3.3 Association between emotional descriptors

We obtain a mapping between annotated primary emotions, affective phenomena, and activation levels by evaluating their associations. This created a comprehensive understanding of their co-occurrence and validated the annotation scheme by demonstrating consistency between the labels of the emotional descriptors.

The Fisher exact tests examined the overall association between the categorical variables: primary emotions with affective phenomena, primary emotions and activation levels, and lastly affective phenomena with activation levels. All obtained a p-value (p) ≦ 0.001 after 10000 simulations with replicates. Each emotional descriptor had more than two labels, so the OR could not be calculated. Thus, the pairwise Fisher exact tests with multiple comparison correction mapped whether two labels were significantly associated with each other, with their OR indicating if there is no association (OR = 1), a positive association (OR > 1) or a negative association (OR < 1) (Figure 3).

Figure 3
Three heatmaps display the relationship between primary emotions, affective phenomena, and activation levels. Heatmap A shows how different emotions correspond to affective phenomena. Heatmap B illustrates the correlation between primary emotions and activation levels. Heatmap C combines activation levels with affective phenomena. Color coding indicates odds ratios, with blue representing lower odds and red higher odds. Each map contains annotated significance indicators.

Figure 3. Heatmap of adjusted p-values of pairwise Fisher’s exact tests with Bonferroni multiple comparison correction. (A) Mapping between primary emotions and affective phenomena labels; (B) Mapping between primary emotions and activation levels labels; (C) Mapping between activation levels and affective phenomena labels. Significant adjusted p-values are displayed by *,**,*** corresponding to p-values equal to 0.05, 0.005, ≤0.001.

As expected, there was a strong positive association of the primary emotion neutral with the affective phenomena neutral and activation level 0 (p ≦ 0.001, OR ≧ 20). Similarly, the affective phenomena consistently co-occurred with their related primary emotions that constituted a lower characterization of emotional experience, relating as a foundational component for more complex affective behaviors that are affective phenomena. The strength of positive associations of primary emotion sad was strong with affective phenomena sad, depressed, hopeless, distressed (p ≦ 0.001, OR ≧ 20), mild for pessimistic and worried (p ≦ 0.001, OR between 2 and 10), and weak for nervous (p ≦ 0.05, OR between 2 and 10). The mapping with angry displayed a strong positive association with irritable, quick tempered, and infuriating (p ≦ 0.001, OR ≧ 20), and mild for frustrated (p ≦ 0.001, OR between 10 and 20). The primary emotion other was strongly and positively associated with disoriented, and confused (p ≦ 0.001, OR ≧ 20), mildly with other (affective phenomena) and weakly with apathetic (p ≦ 0.001, OR between 2-10). Lastly, As the PBA assessed psychiatric symptoms and does not have any positive valence, happy was only associated positively with the affective phenomena other (p ≦ 0.001, OR ≧ 20).

Activation levels equal to 1 were mildly and positively associated with the primary emotion other (p ≦ 0.001, OR between 2-10). Regarding activation levels 2 and 3, mild positive associations were found with sad, angry, and happy (p ≦ 0.001, OR between 2-10).

Lastly, the mapping displayed multiple negative associations between the labels, which illustrated that they do not occur simultaneously in the emoHD corpus (OR between 0 and 0.80) (for example the primary emotion happy does not occur with the affective phenomena irritable (p ≦ 0.001, OR ≦ 1).

3.4 Emotional expression in Huntington’s disease gene carriers and healthy controls

3.4.1 Primary emotions

HD gene carriers and HC differed regarding their proportion of expressed angry (t-statistic (t) = 3.46 (51.12), p = 0.001, permutation p-value (perm-p) ≦ 0.001) with HC exhibiting higher ((x¯) = 0.165) proportions angry than HD (x¯ = 0.103) (Figure 4). Tukey’s test revealed that this was carried by gene carriers in the HD ISS stage 3 exhibiting a lower proportion than the HC individuals (estimate = -0.07, p ≤ 0.001). The remaining proportions of primary emotions were similar across HD gene carriers and HC (all p ≧ 0.084, perm-p ≧ 0.108): proportion of sad (t = -1.75 (83.07)), happy (t = -1.50 (84.03)), neutral (t = -0.47 (75.26)) and other (t = -1.54 (70.67)). See Supplementary Table 3 for details.

Figure 4
Violin plots comparing groups HC and HD across five emotion categories: Sad, Angry, Happy, Neutral, and Other. A significant difference in the Angry category is indicated by asterisks and a bar. Green represents HC, and yellow represents HD.

Figure 4. Proportion of primary emotions in Huntington’s disease gene carriers and controls. Healthy controls (HC) and Huntington’s disease (HD) gene carriers’ distributions of primary emotions proportions. The proportions range from 0-1. Significant group differences are displayed by *,**,*** corresponding to permutation p-values equal to 0.05, 0.005, ≤0.001.

3.4.2 Affective phenomena

The emoHD corpus participants expressed some affective phenomena differently (Figure 5). Pairwise group difference assessed by the Tukey’s test based on the HD ISS revealed which groups carried the difference.

Figure 5
Violin plots comparing emotional states between two groups, HC and HD, using green and yellow colors, respectively. Emotional states like apathetic, confused, and irritable show significant differences marked by asterisks.

Figure 5. Proportions of affective phenomena in Huntington’s disease gene carriers and controls. Healthy controls (HC) and Huntington’s disease (HD) gene carriers’ distributions of affective phenomena (AP) proportions. The proportions range from 0-1. Significant group differences are displayed by *,**,*** corresponding to permutation p-values equal to 0.05, 0.005, ≤0.001.

The proportion of apathetic (HD x¯ = 0.024, HC = x¯ = 0.001, t = -3.16 (77.22), p = 0.002, perm-p = 0.025) and depressed (HD x¯ = 0.007, HC x¯ = 0.000, t = -2.82 (76.25), p = 0.006, perm-p = 0.049) were higher in HD gene carriers than HC, with no specific pairwise group differences (all p ≧ 0.63).

The proportions of confused (HD x¯ = 0.043, HC x¯ = 0.009, t = -4.17 (99.62), p ≦ 0.001, perm-p = 0.003), disoriented (HD x¯ = 0.011, HC x¯ = 0.002, t = - 3.50 (102.32), p ≦ 0.001, perm-p = 0.011) were higher in HD gene carriers than HC: HD ISS 3 having higher proportions than HC (confused estimate = 0.040, p = 0.004; disoriented estimate = 0.001, p = 0.041). Pessimistic was also higher in HD (HD x¯ = 0.012, HC x¯ = 0.002, t = - 3.20 (99.32), p = 0.002, perm-p = 0.020), HD ISS 2 having higher proportions than HC (estimate = 0.02, p = 0.033). Proportions of expressed frustrated were higher in HD than HC (HD x¯ = 0.048, HC x¯ = 0.025, t = - 2.70 (100.03), p = 0.008, perm-p = 0.022) due to HD ISS 0–1 carrying these higher proportions compared of HC (estimate= 0.04, p = 0.018).

Proportions of other (HD = 0.295, HC = 0.360, t = 2.48 (78.09), p = 0.016, perm-p = 0.021) and irritable (HD = 0.085, HC = 0.121, t = 2.08 (50.58), p = 0.043, perm-p = 0.020) were lower in the HD group than HC. HC had higher proportions than the HD ISS 3 for both these affective phenomena (other estimate = - 0.08, p = 0.05; irritable estimate = - 0.04, p = 0.394).

No other affective phenomena displayed a group effect when comparing HD gene carriers and HC (all p & perm-p ≧ 0.05). See Supplementary Table 4 for details.

3.4.3 Activation

HD gene carriers showed higher proportions of activation levels equal to 0 than HC (HD = 0.313, HC = 0.203, t = -3.59 (64.22), p & perm-p ≦ 0.001), carried by HD ISS 3 and 2 having higher proportions than HC (respectively estimate = 0.12, p = 0.002; estimate = 0.12, p = 0.138) (Figure 6). HD gene carriers exhibited lower proportions of activation levels equal to 1 than HC (HD = 0.472, HC = 0.530, t = 2.09 (71.56), p = 0.040, perm-p = 0.045), with no specific pairwise differences between HD ISS groups and controls (p ≧ 0.148). There were no other differences in activation levels (p ≧ 0.144, perm-p ≧ 0.126): activation equal to 2 with t = 1.47 (59.52) and equal to 3 with t = 1.19 (47.06). See Supplementary Table 5 for details.

Figure 6
Violin plots showing data distributions for two groups, HC (green) and HD (yellow), across four activation levels (0 to 3). Significant differences are noted for activations 0 and 1, indicated by asterisks above the plots.

Figure 6. Proportion of activation levels in Huntington’s disease gene carriers and controls. Healthy control (HC) and Huntington’s disease (HD) gene carriers’ distributions of activation levels proportions. The proportions range from 0-1. Significant group differences are displayed by *,**,*** corresponding to permutation p-values equal to 0.05, 0.005, ≤0.001.

3.5 Associations between emotions and psychiatric behaviors

Multilinear regressions evaluated the associations between the proportions of emotional descriptor labels with each PBA sub-scale score (Figure 7). Using a stepwise bidirectional approach, we determined which important emotional descriptor labels were associated with psychiatric scores, adjusting for sex, age and the clinical group status (HD ISS 0-1, 2, 3 and HC).

Figure 7
Plots depicting estimates for different psychological phenomena: Depression, Apathy, Irritability/Aggressivity, and Obsessive/Compulsive. Each plot shows horizontal estimates with significance levels. Categories include activation levels, affective phenomena, primary emotions, and HD ISS factors. Significance is indicated with p-values and color coding: dark circles for p ≤ 0.05 and open circles for p > 0.05. Each psychological phenomenon is broken down into specific attributes, such as “Nervous” or “Aphetic”, with corresponding p-values next to estimates.

Figure 7. Multilinear regressions after a stepwise bidirectional approach for PBA sub-scales. Healthy controls (HC) and Huntington’s Disease (HD) gene carriers grouped according to the Integrating Staging System (ISS). The figure displays the results of the multilinear regressions with stepwise bidirectional approach, evaluating the associations between the different emotional descriptors (primary emotions, affective phenomena, and activation levels) and the Depression, Irritability/aggressivity, Apathy and Obsessive/compulsive Problem Behaviors Assessment (PBA) sub-scales. An increase in PBA sub-scales characterizes worst psychiatric symptoms. The Depression sub-scale ranges from 0 to 48, Irritability/aggressivity and Obsessive/compulsive sub-scales from 0 to 32, and Apathy from 0 to 16. The emotional descriptors are expressed in proportions normalizing total speech duration. The bidirectional stepwise approach selected the emotional descriptors that best fit the PBA sub-scales, and displaying the significantly associated ones by a filled dot and by *,**,*** corresponding to p-values equal to 0.05, 0.005, ≤0.001.

The expression of affective phenomena anxious (p = 0.001, confident interval (CI) = 46.63 – 195.73, estimate = 121.18) and nervous (p = 0.010, CI = 28.53 – 206.28, estimate = 117.40) were positively associated to the Depression sub-scale (R2 = 0.317). In other words, greater expression of these affective phenomena is associated with a higher Depression sub-scale score. On the other hand, higher expression of the affective phenomena disoriented appeared to be preventive for Depression (p = 0.031, CI = -117.72 – -6.32, estimate = -62.02). Depression appeared slightly associated with HD ISS 3 (p = 0.004, CI = 1.14 – 5.85, estimate = 3.49).

Irritability/aggressivity sub-scale (R2 = 0.190) was positively associated with the affective phenomenon quick tempered and frustrated (respectfully p = 0.045, CI = 0.51 – 28.87, estimate = 14.69; and p = 0.037, CI = 0.42-13.87, estimate = 7.14), and the primary emotion sad (p = 0.039, CI = 0.30 – 9.39, estimate = 4.84) without any association with the HD ISS status.

The Apathy sub-scale (R2 = 0.414) was strongly and positively associated with the affective phenomena pessimistic (p = 0.001, CI = 11.15 – 43.25, estimate = 27.20), followed by depressed (p = 0.010, CI = 5.73 – 42.62, estimate = 24.17), apathetic (p ≦ 0.001, CI = 9.29 – 22.69, estimate = 15.99) and hopeless (p = 0.001, CI = 1.97 – 22.46, estimate = 12.22), and weakly with the primary emotion neutral (p = 0.040, CI = 0.17 – 7.32, estimate = 3.75). No effect of the HD ISS status was identified.

The Obsessive/compulsive sub-scale (R2 = 0.512) was associated positively with the affective phenomena fearful (p ≦ 0.001, CI = 126.44 – 233.37, estimate = 179.90) followed by apathetic (p = 0.001, CI = 5.58 – 20.11, estimate = 12.84), and with the HD ISS stage 3 (p ≦ 0.001, CI = 1.45 – 3.26, estimates = 2.35).

Across the four multilinear models, we observed weak to moderate negative associations between the scales and the activation levels 0, 1, and/or 2 (p ≦ 0.05) indicating that individuals with higher psychiatric symptoms tended to exhibit lower excitation levels in their voice. We also see that the highest level of activation (equal to 3) was not associated with any sub-scale. Detailed results are available in the Supplementary Table 69.

4 Discussion

We explored the connection between emotional expression and psychiatric symptoms in HD by developing the first annotated HD emotional speech corpus, emoHD, and assessed its congruence with the PBA. By comparing HD gene carriers and HC, we found that HD individuals exhibited slightly more Depression, Apathy and Obsessive/compulsive symptoms on the PBA sub-scales, but surprisingly similar to controls regarding Irritability/aggressivity. HD speech contained a greater proportion of segments with no emotional activation (level 0), while fewer segments with mild emotional activation (level 1), denoting fewer emotional expressions than controls. HD gene carriers and controls did not differ regarding their expression of primary emotions, except for a lower expression of angry in HD, primarily due to patients in later disease stages. In contrast, HD gene carriers expressed more the affective phenomena apathetic, confused, depressed, disoriented, frustrated, and pessimistic than controls, while expressing less other and irritable. This suggests that the affective phenomena framework is more appropriate to capture emotional speech in HD in comparison to primary emotions. Altogether, we demonstrated for the first time, a congruence between the emotional expressions in speech and psychiatric symptoms in HD, revealing which specific emotional expressions are indicative of an increase in the PBA sub-scale scores (Table 5).

Table 5
www.frontiersin.org

Table 5. Summary of main findings.

One might be surprised that angry was less expressed in HD patients’ speech than in controls. However, previous research on the same cohort has shown that, when assessing the general emotional content of emotional speech tasks, patients narrating anger and happy stories exhibited lower emotional content compared to healthy individuals (67). This lower expression of angry aligns with HD’s consistently reported difficulty in recognizing facial and vocal emotional expressions of anger (6871), as well as the observed tendency to be less capable of facially imitating anger when prompted by a card (72). Our analysis of emotional segments of HD patient’s speech, showed that they not only struggle in recalling and narrating anger-charged stories but also express that emotion less. Furthermore, in our study, patients spoke more without emotional activation (level 0) and less at mild emotional activation (level 1) than controls, showing a general diminished capacity for oral emotional expression. This might be linked to the dysprosody in HD where patients may struggle to convey emotions through vocal tone, pitch and prosodic variation (5, 73, 74). Patients’ dysarthria, a motor disorder affecting speech clarity, articulation, vocal control and speech rate (75), may also contribute to decreasing the effectiveness of their communication. However, research on patients without buccofacial motor disorders suggests that oral and facial emotional production deficits are not solely due to motor impairments, but also involve complex cognitive, neuropsychiatric and emotional processing deficits (3, 7, 72, 75).

In contrast, patients spontaneously expressed many emotions not elicited by the emotional-speech tasks (e.g. sad, angry and happy), captured under the emotional descriptor affective phenomena. These were more nuanced than Ekman’s universal emotions, which have been criticized for being too rigid and simplistic to capture the complexity of real-life emotional experiences (76). We addressed this challenge by selecting the affective phenomena labels as listed by external participants exposed to PBA statements, to minimize potential clinical and research bias. Unlike other emotional speech corpora featuring pathological speakers (21, 22, 37, 39), we focus on emotional states reflecting potential disease-specific manifestations, rather than on basic, coarse-grained emotions. In this context, affective phenomena provided a clearer way to identify predominant emotional behaviors in HD such as confusion, disorientated, and pessimistic. These became more pronounced as the disease progressed, particularly in HD ISS 2 or 3, which is congruent with the documented increase in obsessive-compulsive and disorientation PBA item scores across HD stages (77). Apathetic and depressed expressions were similar across disease stages, suggesting that these may be experienced throughout the disease course as affective dispositions or moods [i.e., personality disposition or an episodic state as defined by Scherer (43)]. This was the case regardless of a formal diagnosis of depression or apathy which occurred in later disease stages in the emoHD cohort. We further found that gene carriers without any motor or cognitive symptoms (ISS 0-1) expressed more frustrated, presumably due to the anticipated disease progression and the negative impact of a recent diagnosis. This negative impact has been shown to affect multiple psychiatric symptoms in HD, even before onset of the disease (78). Surprisingly, HD patients exhibited less irritability than controls, consistent with the lack of group difference on the Irritability/aggressivity sub-scale. One possibility is that patients were genuinely not irritable at the time of the interview, as those experiencing irritability may have opted out of the speech recordings. Another possibility is that patients’ behavior in the hospital setting may differ from their behavior at home.

Limited research has explored the interplay between emotions and psychiatric behaviors in HD. Existing studies consistently reported associations between apathy - the most prevalent psychiatric symptom in HD (79) - and impaired facial emotion recognition, typically assessed using tasks that evaluate Ekman’s emotions (8082). Our study included primary emotions but extended the scope of investigation by introducing affective phenomena and examining their relationships with a wider range of psychiatric symptoms. Alternative approaches have linked HD psychiatric symptoms to self-reported emotional well-being assessed with quality of life scales, showing that lower emotional well-being correlated with more severe symptoms (83, 84). However, quality of life scales may limit the set of emotional questions, restricting patients’ from naturally expressing their emotions and feelings, and therefore failing to accurately capture their emotional state (85). Spontaneous speech is a more ecologically valid method because it can directly link the expressed emotions cross-sectionally with the patient’s current psychiatric well-being.

Although corpora from healthy individuals have been used to develop speech emotion recognition models for conditions like Alzheimer’s and Parkinson’s disease (3941), neurological disorders include unique language, speech, cognitive and emotional deficits, thus, making the use of healthy individuals’ data insufficient to fully represent the profiles of the targeted neurological conditions (86). The emoHD corpus followed the principles outlined by Douglas-Cowie for creating a high-quality and generalizable emotional speech corpus (87): scope of speakers, naturalness of speech, context and emotional descriptors. Indeed, EmoHD targeted a large sample of genetically confirmed HD gene carriers with varied disease clinical stages, alongside control participants for comparison purposes. We prioritized elicited emotional speech to maximize the ecological validity of the corpus, having a trade-off between the limitations of acted emotional speech — which fails to convey real-life emotions (88) — and natural emotional databases — which often face legal and ethical issues (89, 90). The emotional speech tasks successfully elicited primary emotions, as evidenced by their predominance in their respective tasks, and allowed for the expression of non-elicited emotions, captured through affective phenomena (Figure 2). Regarding contextual information, our design included psychiatric assessments conducted by certified neurologists — an aspect often lacking in emotional speech corpora with neurological patients as speakers. Indeed, such corpora prioritize cognitive or motor assessments over psychiatric insights (36, 38, 90, 91), or rely on self-reported psychiatric scales that may not accurately reflect the patient’s psychiatric state (35, 37, 92). Lastly, by annotating emotions and quantifying their duration within patients’ speech, we captured emotion-carrying segments of speech. In contrast, prior work on HD emotional expression (5, 67) typically assumed that the emotion elicited remained present throughout the entire recording, using it as the emotional label. While eliciting emotions (93) and successful narration must arouse emotion in speakers (55, 56), these approaches implied that emotional expression remains constant throughout the narrative speech— which was rarely the case in the annotated emoHD corpus.

Speech appears as a non-invasive marker of HD patients’ emotional well-being and psychiatric status, complementing its established use in the assessment of motor, functional and cognitive performance (52, 9499). More generally, speech has been recognized as an insightful marker for psychological and psychiatric monitoring. Alteration in speech signals (e.g., prosodic, rhythm changes, etc.) (28, 29, 100) and lexical and semantic patterns (101) are markers of various conditions. This includes mood disorders such as depression and anxiety, psychotic disorders, or emotional processing deficits like alexithymia. Using speech as a non-invasive objective marker has encouraged the development of machine learning pipelines that analyze speech to detect and monitor mental health conditions (20, 28). By analyzing an individual’s speech frequently, these models could help clinicians adapt personalized treatment plans, detect early warning signs, and deliver targeted interventions based on speech-derived markers.

The annotation and validation of the emoHD dataset in our study makes it now possible to use it to train speech emotion recognition (SER) models. These models are designed to identify expressed emotions based on the speakers’ acoustic signal (102). In the case of HD, a SER model could enable the automatic annotation of patients’ speech, providing a scalable tool to detect emotional disturbances. This output can then be further used to predict their psychiatric behaviors. As our speech recordings ranged from 4.3 to 5 minutes, this approach would be notably less time-consuming than a full PBA evaluation, which typically lasts around 20 mins (103). Hence integrating a SER model into HD clinical care would support the remote monitoring of emotional and psychiatric symptoms. This would also assist in the early detection of behavioral changes and track the progression of these symptoms and thus enhance personalized medicine. Given the challenge many HD patients face in accessing healthcare systems — such as medical experts, medical deserts and long appointment delays (104106) — speech, which is easy to collect, would allow to perform additional clinical assessment of patients’ in their homes, in-between physical medical appointments.

Some limitations and challenges should be considered in future work. First, although we included 102 HD gene carriers, the sample size remains restrained for normative data in a neurodegenerative disease. Nevertheless, our sample size remains acceptable for a rare disease as HD, and our cohort is larger than most existing emotional speech corpus with pathological speakers (35, 41), providing a valuable contribution in a field where datasets are often limited. As emoHD is an elicited emotional dataset, the recording conditions ensured consistency across participants, minimizing background noise and language-related confounds. This allowed us to reliably isolate disease-specific emotional markers of psychiatric symptoms. Second, in the present study, emotional annotation remained a time-consuming process (30 minutes per recording). However, now that the emoHD corpus has been annotated, it can be used to train a SER model, enabling future recordings to be automatically annotated without the need for manual annotations. Third, while we were able to explore the associations between emotional expressions and psychiatric behaviors, our cohort suffers the same bias as all HD cohorts, namely not including individuals at high-risk of engaging in harmful behaviors towards themselves or others, as urgent care is of higher priority than volunteering to partake in studies. Apart from the Irritability/aggressivity sub-scales that did not differ between HD ISS and controls, the scores on the PBA sub-scales remained similar to other cohorts like REGISTRY HD (107). To account for these limitations, we enhanced our statistical power by applying permutation based statistical tests (63, 64), and incorporating healthy controls with some psychiatric symptoms. Additionally, post-hoc analysis of possible cofounders of speech have been evaluated (Supplementary Table 10), suggesting that language cognitive abilities, educational level and antipsychotic drugs influence HD participants total speech time, which should be taken into consideration in future studies. Lastly, regarding the valence of our different emotional descriptors’ labels, we are aware that they are majorly negative. Indeed, the affective phenomena were chosen based on external participants’ interpretations of the PBA, which lacks any positive statements. We anticipated that the primary emotion happy would either not show any associations, negatively correlate with most affective phenomena or be positively associated with the label other, which was the case when we analyzed the emotional descriptors associations. As our aim is not to evaluate patients’ quality of life or positivity, we acknowledge this limitation, which does not significantly impact our approach in the exploration of emotions and their associations to psychiatric behaviors.

5 Conclusion

By developing the emoHD corpus, we were able to associate emotional expression and psychiatric behaviors in Huntington’s Disease underscoring the relevance of our findings in this population. This work encourages future work to specifically develop speech-based digital tools that can assess Huntington’s Disease patients emotional and psychiatric well-being. HD serves as an excellent model for studying neurodegenerative disorders (108) because of the multifaceted nature of motor, cognitive, and psychiatric symptoms and their interplay. We further propose that the emoHD framework’s flexibility and adaptability can be developed and applied in other neurodegenerative diseases with affective disorders, by applying the proposed method to the collected speech samples. This would identify other disease-specific emotional and psychiatric associations.

Data availability statement

The conditions of our ethics approval, which includes the ethical consent of participants and the collected data, do not allow for public archiving of anonymized study data. On reasonable request, readers seeking access to the data should contact the corresponding author. Access will be granted to named individuals in accordance with ethical procedures for the reuse of sensitive data, including a research partnership and completion of a data transfer partnership by the APHP. Legal copyright restrictions prevent public archiving of the UHDRS, which can be obtained from UHDRS® | - Huntington Study Group, which can be obtained from the agency pour la protection des programmes IDDN FR 001 230010 001 S A 2020 000 31 230. Statistical analysis script can be found on https://gitlab.cognitive-ml.fr/lchenain/emohd_corpus.git.

Ethics statement

Ethics approval was given by the institutional review board from Henri Mondor Hospital (Créteil, France) for the BIO-HD cohort and the CPP Saint Louis French part of the REPAIR-HD cohort, in compliance with the Helsinki Declaration. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

LC: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Resources, Software, Visualization, Writing – original draft, Writing – review & editing. AF: Resources, Writing – review & editing. HT: Methodology, Resources, Software, Writing – review & editing. GM: Writing – review & editing. KY: Investigation, Resources, Writing – review & editing. CC: Conceptualization, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Validation, Writing – original draft, Writing – review & editing. A-CB-L: Conceptualization, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Validation, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research and/or publication of this article. This study was funded by Repair-HD from the European Union’s Seventh Framework Programme; for research, technological development and demonstration under grant agreement n°602245 (www.repair-hd-eu); Bio-HD by the Henri-Mondor Hospital National Reference Centre for Huntington’s disease (Ministry of Health); Agence Nationale pour la Recherche (ANR-11-INBS-0011 NeurATRIS; ANR-17-EURE-0017 Frontcog; ANR-10-IDEX-0001-02 PSL*; ANR-23-CE23-0033-01 SINNet; ANR-23-IACL-0008 France 2030 PRAIRIE). LC is a student from the FIRE PhD program, funded by the Bettencourt Schueller foundation and the EURIP graduate program (ANR-17-EURE-0012).

Acknowledgments

We are very thankful and grateful for the participants that accepted to take place in our study. We thank the speech therapists that contributed to the annotation of our data. Furthermore, we thank Andres Gil-Salcedo, the NPI team, the National Reference Centre for Huntington’s Disease (Ministry of Health), the CeReNe team at the department of neurology at the Henri Mondor Hospital for their help to carry out this project, as well as the ALMANACH team.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyt.2025.1633492/full#supplementary-material

References

1. Bates GP, Dorsey R, Gusella JF, Hayden MR, Kay C, Leavitt BR, et al. Huntington disease. Nat Rev Dis Primers. (2015) 1:15005. doi: 10.1038/nrdp.2015.5

PubMed Abstract | Crossref Full Text | Google Scholar

2. Walker FO. Huntington’s disease. Lancet. (2007) 369:218–28. doi: 10.1016/S0140-6736(07)60111-1

PubMed Abstract | Crossref Full Text | Google Scholar

3. Trinkler I, Devignevielle S, Achaibou A, Ligneul RV, Brugières P, de Langavant LC, et al. Embodied emotion impairment in Huntington’s Disease. Cortex. (2017) 92:44–56. doi: 10.1016/j.cortex.2017.02.019

PubMed Abstract | Crossref Full Text | Google Scholar

4. Hayes CJ, Stevenson RJ, and Coltheart M. Production of spontaneous and posed facial expressions in patients with Huntington’s disease: Impaired communication of disgust. Cogn Emotion. (2009) 23:118–34. doi: 10.1080/02699930801949090

Crossref Full Text | Google Scholar

5. Gallezot C, Riad R, Titeux H, Lemoine L, Montillot J, Sliwinski A, et al. Emotion expression through spoken language in Huntington disease. Cortex. (2022) 155:155. doi: 10.1016/j.cortex.2022.05.024

PubMed Abstract | Crossref Full Text | Google Scholar

6. Bora E, Velakoulis D, and Walterfang M. Social cognition in Huntington’s disease: A meta-analysis. Behav Brain Res SreeTestContent1. (2016) 297:131–40. doi: 10.1016/j.bbr.2015.10.001

PubMed Abstract | Crossref Full Text | Google Scholar

7. Kordsachia CC, Labuschagne I, and Stout JC. Beyond emotion recognition deficits: A theory guided analysis of emotion processing in Huntington’s disease. Neurosci Biobehav Rev. (2017) 73:276–92. doi: 10.1016/j.neubiorev.2016.11.020

PubMed Abstract | Crossref Full Text | Google Scholar

8. Eddy CM, Parkinson EG, and Rickards HE. Changes in mental state and behaviour in Huntington’s disease. Lancet Psychiatry. (2016) 3:1079–86. doi: 10.1016/S2215-0366(16)30144-4

PubMed Abstract | Crossref Full Text | Google Scholar

9. Erlangsen A, Stenager E, Conwell Y, Andersen PK, Hawton K, Benros ME, et al. Association between neurological disorders and death by suicide in Denmark. JAMA. (2020) 323:444–54. doi: 10.1001/jama.2019.21834

PubMed Abstract | Crossref Full Text | Google Scholar

10. Paulsen JS, Langbehn DR, Stout JC, Aylward E, Ross CA, Nance M, et al. Detection of Huntington’s disease decades before diagnosis: the Predict-HD study. J Neurol Neurosurg Psychiatry. (2008) 79:874–80. doi: 10.1136/jnnp.2007.128728

PubMed Abstract | Crossref Full Text | Google Scholar

11. Duff K, Paulsen JS, Beglinger LJ, Langbehn DR, and Stout JC. Psychiatric Symptoms in Huntington’s Disease before Diagnosis: The Predict-HD Study. Biol Psychiatry. (2007) 62:1341–6. Available online at: https://www.sciencedirect.com/science/article/pii/S0006322306015617 (Accessed November 04, 2023).

Google Scholar

12. Marder K, Zhao H, Myers RH, Cudkowicz M, Kayson E, Kieburtz K, et al. Rate of functional decline in Huntington’s disease. Huntington Study Group. Neurology. (2000) 54:452–8. doi: 10.1212/WNL.54.2.452

PubMed Abstract | Crossref Full Text | Google Scholar

13. Helder DI, Kaptein AA, van Kempen GM, van Houwelingen JC, and Roos RA. Impact of Huntington’s disease on quality of life. Mov Disord. (2001) 16:325–30. doi: 10.1002/mds.1056

PubMed Abstract | Crossref Full Text | Google Scholar

14. Ho AK, Robbins AOG, Walters SJ, Kaptoge S, Sahakian BJ, and Barker RA. Health-related quality of life in Huntington’s disease: A comparison of two generic instruments, SF-36 and SIP. Movement Disord. (2004) 19:1341–8. doi: 10.1002/mds.20208

PubMed Abstract | Crossref Full Text | Google Scholar

15. Ho AK, Gilbert AS, Mason SL, Goodman AO, and Barker RA. Health-related quality of life in Huntington’s disease: Which factors matter most? Movement Disord. (2009) 24:574–8. doi: 10.1002/mds.22412

PubMed Abstract | Crossref Full Text | Google Scholar

16. Ready RE, Mathews M, Leserman A, and Paulsen JS. Patient and caregiver quality of life in Huntington’s disease. Movement Disord. (2008) 23:721–6. doi: 10.1002/mds.21920

PubMed Abstract | Crossref Full Text | Google Scholar

17. Junca E, Pino M, Santamaría-García H, and Baez S. Brain, cognitive, and physical disability correlates of decreased quality of life in patients with Huntington’s disease. Qual Life Res. (2023) 32:171–82. doi: 10.1007/s11136-022-03220-0

PubMed Abstract | Crossref Full Text | Google Scholar

18. Youssov K, Audureau E, Vandendriessche H, Morgado G, Layese R, Goizet C, et al. The burden of Huntington’s disease: A prospective longitudinal study of patient/caregiver pairs. Parkinsonism Relat Disord. (2022) 103:77–84. Available online at: https://www.sciencedirect.com/science/article/pii/S1353802022002802 (Accessed February 11, 2024).

PubMed Abstract | Google Scholar

19. Bachoud-Lévi AC, Ferreira J, Massart R, Youssov K, Rosser A, Busse M, et al. International guidelines for the treatment of huntington’s disease. Front Neurol. (2019) 10:710/full. doi: 10.3389/fneur.2019.00710/full

PubMed Abstract | Crossref Full Text | Google Scholar

20. Cummins N, Baird A, and Schuller BW. Speech analysis for health: Current state-of-the-art and the increasing impact of deep learning. Methods. (2018) 151:41–54. Available online at: https://www.sciencedirect.com/science/article/pii/S1046202317303717 (Accessed July 31, 2024).

PubMed Abstract | Google Scholar

21. Ringeval F, Schuller B, Valstar M, Cummins Ni, Cowie R, Tavabi L, et al. AVEC 2019 workshop and challenge: state-of-mind, detecting depression with AI, and cross-cultural affect recognition. arXiv preprint. (2019). Available online at: http://arxiv.org/abs/1907.11510 (Accessed December 23, 2024).

Google Scholar

22. de Angel V, Lewis S, Munir S, Matcham F, Dobson R, and Hotopf M. Using digital health tools for the Remote Assessment of Treatment Prognosis in Depression (RAPID): a study protocol for a feasibility study. BMJ Open. (2022) 12:e059258. doi: 10.1136/bmjopen-2021-059258

PubMed Abstract | Crossref Full Text | Google Scholar

23. Tasnim M, Ehghaghi M, Diep B, and Novikova J. DEPAC: a corpus for depression and anxiety detection from speech. arXiv preprint. (2023). Available online at: http://arxiv.org/abs/2306.12443 (Accessed December 16, 2024).

Google Scholar

24. Faurholt-Jepsen M, Busk J, Frost M, Vinberg M, Christensen EM, Winther O, et al. Voice analysis as an objective state marker in bipolar disorder. Transl Psychiatry. (2016) 6:e856. doi: 10.1038/tp.2016.123

PubMed Abstract | Crossref Full Text | Google Scholar

25. Quillivic R, Gayraud F, Auxéméry Y, Vanni L, Peschanski D, Eustache F, et al. Interdisciplinary approach to identify language markers for post-traumatic stress disorder using machine learning and deep learning. Sci Rep. (2024) 14:12468. Available online at: https://www.nature.com/articles/s41598-024-61557-7 (Accessed January 07, 2025).

PubMed Abstract | Google Scholar

26. Stasak B, Epps J, Schatten HT, Miller IW, Provost EM, and Armey MF. Read speech voice quality and disfluency in individuals with recent suicidal ideation or suicide attempt. Speech Communication. (2021) 132:10–20. Available online at: https://www.sciencedirect.com/science/article/pii/S0167639321000546 (Accessed December 23, 2024).

Google Scholar

27. Scherer S, Pestian J, and Morency LP. Investigating the speech characteristics of suicidal adolescents. In: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (2013). p. 709–13. Available online at: https://ieeexplore.ieee.org/abstract/document/6637740?casa_token=7J0rHzUDCdEAAAAA:ed5OEhIG4zDbDQ95_fNebBKKWxqGEw_aP9QJmNLtzWAlEvV0RbJQyiM5M7Zni0-6m-GZNbUgkU0.

Google Scholar

28. Low DM, Bentley KH, and Ghosh SS. Automated assessment of psychiatric disorders using speech: A systematic review. Laryngoscope Invest Otolaryngol. (2020) 5:96–116. Available online at: https://onlinelibrary.wiley.com/doi/abs/10.1002/lio2.354 (Accessed November 14, 2024).

Google Scholar

29. Kappen M, Vanderhasselt MA, and Slavich GM. Speech as a promising biosignal in precision psychiatry. Neurosci Biobehav Rev. (2023) 148:105121. Available online at: https://www.sciencedirect.com/science/article/pii/S0149763423000908 (Accessed December 11, 2024).

PubMed Abstract | Google Scholar

30. Huang KY, Wu CH, Su MH, and Kuo YT. Detecting unipolar and bipolar depressive disorders from elicited speech responses using latent affective structure model. IEEE Trans Affect Comput. (2020) 11:393–404. Available online at: https://ieeexplore.ieee.org/document/8288569?denied (Accessed August 2, 2024).

Google Scholar

31. Dikaios K, Rempel S, Dumpala SH, Oore S, Kiefte M, and Uher R. Applications of speech analysis in psychiatry. Harvard Rev Psychiatry. (2023) 31:1. Available online at: https://journals.lww.com/hrpjournal/fulltext/2023/01000/applications_of_speech_analysis_in_psychiatry.1.aspx?casa_token=hfyVRI3ovdEAAAAA:atlMkvolv_UNQnv8lTgfFo-fonQ9gkDAL8WjIZ1JG6UXqBJHZFp9vhlEyA4b1yImmj2s9czm2GbSACl0VZ0uFBCD1QSzSsvIz-cLHQFvLQU (Accessed December 11, 2024).

Google Scholar

32. Scherer KR. Psychological models of emotion. In: Borod JC, editor. The neuropsychology of emotion. Oxford; New York: Oxford University Press (2000) p. 137–66. (Series in affective science).

Google Scholar

33. Akçay MB and Oğuz K. Speech emotion recognition: Emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers. Speech Communication. (2020) 116:56–76. Available online at: https://www.sciencedirect.com/science/article/pii/S0167639319302262 (Accessed November 11, 2022).

Google Scholar

34. Lopez-de-Ipina K, Martinez-de-Lizarduy U, Calvo PM, Mekyska J, Beitia B, Barroso N, et al. Advances on automatic speech analysis for early detection of alzheimer disease: A non-linear multi-task approach. Curr Alzheimer Res. (2018) 15:139–48. doi: 10.2174/1567205014666171120143800

PubMed Abstract | Crossref Full Text | Google Scholar

35. Zhu T, Duan S, Liang H, and Zhang W. Design, construction and evaluation of emotional multimodal pathological speech database. arXiv preprint. (2023). Available online at: http://arxiv.org/abs/2312.08998 (Accessed December 11, 2024).

Google Scholar

36. Milling M, Baird A, Bartl-Pokorny KD, Liu S, Alcorn AM, Shen J, et al. Evaluating the impact of voice activity detection on speech emotion recognition for autistic children. Front Comput Sci. (2022) 4. Available online at: https://www.frontiersin.org/journals/computer-science/articles/10.3389/fcomp.2022.837269/full (Accessed December 23, 2024).

Google Scholar

37. Zhao S, Rudzicz F, Carvalho LG, Marquez-Chin C, and Livingstone S. Automatic detection of expressed emotion in Parkinson’s Disease. In: 2014 IEEE international Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, Florence, Italy (2014). p. 4813–7. Available online at: http://ieeexplore.ieee.org/document/6854516/ (Accessed November 17, 2022).

Google Scholar

38. Han KH, Zaytseva Y, Bao Y, Pöppel E, Chung SY, Kim JW, et al. Impairment of vocal expression of negative emotions in patients with alzheimer’s disease. Front Aging Neurosci. (2014) 6. Available online at: https://www.frontiersin.org/journals/aging-neuroscience/articles/10.3389/fnagi.2014.00101/full (Accessed January 07, 2025).

PubMed Abstract | Google Scholar

39. Haider F, de la Fuente Garcia S, Albert P, and Luz S. Affective speech for alzheimer’s dementia recognition. (2020).

Google Scholar

40. Bernieri G and Duarte JC. Identificação da Doença de Alzheimer Através da Fala Utilizando Reconhecimento de Emoções. J Health Inf. (2023) 15. Available online at: https://jhi.sbis.org.br/index.php/jhi-sbis/article/view/1093 (Accessed December 16, 2024).

Google Scholar

41. Sechidis K, Fusaroli R, Orozco-Arroyave JR, Wolf D, and Zhang YP. A machine learning perspective on the emotional content of Parkinsonian speech. Artif Intell Med. (2021) 115:102061. doi: 10.1016/j.artmed.2021.102061

PubMed Abstract | Crossref Full Text | Google Scholar

42. Ekman P. Basic emotions. In: Handbook of Cognition and Emotion, 16th ed. San Francisco, USA: John Wiley & Sons (2000). p. 45–60.

Google Scholar

43. Scherer KR. What are emotions? And how can they be measured? Soc Sci Inf. (2005) 44:695–729. doi: 10.1177/0539018405058216

Crossref Full Text | Google Scholar

44. Craufurd D, Thompson JC, and Snowden JS. Behavioral changes in huntington disease. Cogn Behav Neurol. (2001) 14:219. Available online at: https://journals.lww.com/cogbehavneurol/abstract/2001/10000/behavioral_changes_in_huntington_disease.4.aspx (Accessed April 30, 2024).

Google Scholar

45. Osgood C, May W, and Miron M. Cross-cultural Universals of Affective Meaning. University of Illinois Press (1975). Available online at: https://books.google.fr/books?hl=en&lr=&id=Ax9CFVxXYD4C&oi=fnd&pg=PP19&dq=ross-cultural+Universals+of+Affective+Meaning.&ots=-7_Hz3zykR&sig=cQCaRsQb9KeGab-UTAYVG3O8yXAv=onepage&q=ross-cultural%20Universals%20of%20Affective%20Meaning.&f=false (Accessed November 13, 2023).

Google Scholar

46. Kremer HPH and Huntington Study Group X. Unified Huntington’s disease rating scale: reliability and consistency(1996). Available online at: https://repository.ubn.ru.nl/handle/2066/23289 (Accessed April 11, 2023).

Google Scholar

47. Mattis S. Mental Status Examination for organic mental syndrome in the elderly patients. In: Geriatric Psychiatry A Hand Book for Psychiatrics and Primary Care Physicians (1976). Available online at: https://cir.nii.ac.jp/crid/1570009750584096896 (Accessed April 11, 2023).

Google Scholar

48. Tabrizi SJ, Schobel S, Gantman EC, Mansbach A, Borowsky B, Konstantinova P, et al. A biological classification of Huntington’s disease: the Integrated Staging System. Lancet Neurol. (2022) 21:632–44. Available online at: https://www.sciencedirect.com/science/article/pii/S147444222200120X (Accessed October 15, 2024).

PubMed Abstract | Google Scholar

49. Tabrizi SJ, Langbehn DR, Leavitt BR, Roos RA, Durr A, Craufurd D, et al. Biological and clinical manifestations of Huntington’s disease in the longitudinal TRACK-HD study: cross-sectional analysis of baseline data. Lancet Neurol. (2009) 8:791–801. doi: 10.1016/S1474-4422(09)70170-X

PubMed Abstract | Crossref Full Text | Google Scholar

50. Callaghan J, Stopford C, Arran N, Boisse MF, Coleman A, Santos RD, et al. Reliability and factor structure of the short problem behaviors assessment for huntington’s disease (PBA-s) in the TRACK-HD and REGISTRY studies. JNP. (2015) 27:59–64. doi: 10.1176/appi.neuropsych.13070169

PubMed Abstract | Crossref Full Text | Google Scholar

51. Rickards H, De Souza J, van Walsem M, van Duijn E, Simpson SA, Squitieri F, et al. Factor analysis of behavioural symptoms in Huntington’s disease. J Neurol Neurosurg Psychiatry. (2011) 82:411–2. doi: 10.1136/jnnp.2009.181149

PubMed Abstract | Crossref Full Text | Google Scholar

52. Riad R, Lunven M, Titeux H, Cao XN, Hamet Bagnou J, Lemoine L, et al. Predicting clinical scores in Huntington’s disease: a lightweight speech test. J Neurol. (2022) 269:5008–21. doi: 10.1007/s00415-022-11148-1

PubMed Abstract | Crossref Full Text | Google Scholar

53. Riad R. Traitement automatique de la parole et du langage pour un suivi personnalisé de la maladie de Huntington. Université Paris sciences et lettres (2022). Available online at: https://theses.fr/s208350 (Accessed April 24, 2024).

Google Scholar

54. Goodglass H and Kaplan E. Boston Diagnostic Aphasia Examination. Philadelphia: Lea and Febiger (1972).

Google Scholar

55. Willems RM, Nastase SA, and Milivojevic B. Narratives for neuroscience. Trends Neurosci. (2020) 43:271–3. Available online at: https://www.cell.com/trends/neurosciences/abstract/S0166-2236(20)30049-7 (Accessed October 25, 2024).

Google Scholar

56. Habermas T. Emotion and Narrative: Perspectives in Autobiographical Storytelling. Cambridge: Cambridge University Press (2018). Available online at: https://www.cambridge.org/core/books/emotion-and-narrative/51D839F5C9C5D00F2652B2EBE20817DA. (Studies in Emotion and Social Interaction). (Accessed October 25, 2024).

Google Scholar

57. Cummings L. Describing the Cookie Theft picture: Sources of breakdown in Alzheimer’s dementia. Pragmat Soc. (2019) 10:153–76. doi: 10.1075/ps.17011.cum

Crossref Full Text | Google Scholar

58. Boersma P, Weenink D, and van Heuven. V. Praat, a system for doing phonetics by computer. Glot Int. (2001) 5.

Google Scholar

59. Titeux H, Riad R, Cao XN, Hamilakis N, Madden K, Cristia A, et al. Seshat: a tool for managing and verifying annotation campaigns of audio data. In: Proceedings of the Twelfth Language Resources and Evaluation Conference. European Language Resources Association, Marseille, France (2020). p. 6976–82. Available online at: https://aclanthology.org/2020.lrec-1.861 (Accessed April 11, 2023).

Google Scholar

60. Clavel C, Vasilescu I, and Devillers L. Fiction support for realistic portrayals of fear-type emotional manifestations. Comput Speech Lang. (2011) 25:63–83. Available online at: https://www.sciencedirect.com/science/article/pii/S0885230810000240 (Accessed October 25, 2024).

Google Scholar

61. Mathet Y, Widlöcher A, and Métivier JP. The unified and holistic method gamma (γ) for inter-annotator agreement measure and alignment. Comput Linguistics. (2015) 41:437–79. doi: 10.1162/COLI_a_00227

Crossref Full Text | Google Scholar

62. Titeux H and Riad R. pygamma-agreement: Gamma $\gamma$ measure for inter/intra-annotator agreement in Python. J Open Source Softw. (2021) 6:2989. doi: 10.21105/joss.02989

Crossref Full Text | Google Scholar

63. Berry KJ, Johnston JE, and Mielke PW. Nominal-level variables, II. In: The Measurement of Association: A Permutation Statistical Approach, 1st ed. Springer Cham (2018). p. 139–222.

Google Scholar

64. Good PI. Permutation, Parametric, and Bootstrap Tests of Hypotheses (Springer Series in Statistics). 3rd ed. Berlin, Heidelberg: Springer New York, NY (2004) p. 33–66.

Google Scholar

65. Fisher RA. The logic of inductive inference. J R Stat Soc. (1935) 98:39–82. Available online at: https://www.jstor.org/stable/2342435 (Accessed December 19, 2024).

Google Scholar

66. Kim HY. Statistical notes for clinical researchers: Chi-squared test and Fisher’s exact test. Restor Dent Endod. (2017) 42:152–5. Available online at: https://synapse.koreamed.org/articles/1090217 (Accessed November 07, 2024).

PubMed Abstract | Google Scholar

67. Chenain L, Bachoud-Lévi AC, and Clavel C. Acoustic characterization of huntington’s disease emotional expression: an explainable AI approach. In: 12th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos-ACIIW 2024. IEEE, Glasgow, UL (2024). Available online at: https://inria.hal.science/hal-04934178/ (Accessed February 14, 2025).

Google Scholar

68. Sprengelmeyer R, Young AW, Calder AJ, Karnat A, Lange H, Hömberg V, et al. Loss of disgust: Perception of faces and emotions in Huntington’s disease. Brain. (1996) 119:1647–65. doi: 10.1093/brain/119.5.1647

PubMed Abstract | Crossref Full Text | Google Scholar

69. Robotham L, Sauter DA, Bachoud-Lévi AC, and Trinkler I. The impairment of emotion recognition in Huntington’s disease extends to positive emotions. cortex. (2011) 47:880–4. doi: 10.1016/j.cortex.2011.02.014

PubMed Abstract | Crossref Full Text | Google Scholar

70. Calder AJ, Keane J, Young AW, Lawrence AD, Mason S, and Barker RA. The relation between anger and different forms of disgust: implications for emotion recognition impairments in Huntington’s disease. Neuropsychologia. (2010) 48:2719–29. doi: 10.1016/j.neuropsychologia.2010.05.019

PubMed Abstract | Crossref Full Text | Google Scholar

71. Henley SMD, Novak MJU, Frost C, King J, Tabrizi SJ, and Warren JD. Emotion recognition in Huntington’s disease: A systematic review. Neurosci Biobehav Rev. (2012) 36:237–53. doi: 10.1016/j.neubiorev.2011.06.002

PubMed Abstract | Crossref Full Text | Google Scholar

72. Trinkler I, de Langavant LC, and Bachoud-Lévi AC. Joint recognition–expression impairment of facial emotions in Huntington’s disease despite intact understanding of feelings. Cortex. (2013) 49:549–58. doi: 10.1016/j.cortex.2011.12.003

PubMed Abstract | Crossref Full Text | Google Scholar

73. Hertrich I and Ackermann H. Acoustic analysis of speech prosody in Huntington’s and Parkinson’s disease: A preliminary report. Clin Linguistics Phonet. (1993) 7:285–97. doi: 10.1080/02699209308985568

Crossref Full Text | Google Scholar

74. Coulombe V, Joyal M, Martel-Sauvageau V, and Monetta L. Affective prosody disorders in adults with neurological conditions: A scoping review. Int J Lang Commun Disord. (2023) 58:1939–54. doi: 10.1111/1460-6984.12909

PubMed Abstract | Crossref Full Text | Google Scholar

75. Hamilton A, Ferm U, Heemskerk AW, Twiston-Davies R, Matheson KY, Simpson SA, et al. Management of speech, Language and communication difficulties in huntington’s disease. Neurodegenerat Dis Manage. (2012) 2:67–77. doi: 10.2217/nmt.11.78

Crossref Full Text | Google Scholar

76. Coppini S, Lucifora C, Vicario CM, and Gangemi A. Experiments on real-life emotions challenge Ekman’s model. Sci Rep. (2023) 13:9511. Available online at: https://www.nature.com/articles/s41598-023-36201-5 (Accessed February 06, 2025).

PubMed Abstract | Google Scholar

77. Gunn S, Dale M, Ovaska-Stafford N, and Maltby J. Mental health symptoms among those affected by Huntington’s disease: A cross-sectional study. Brain Behav. (2023) 13:e2954. doi: 10.1002/brb3.2954

PubMed Abstract | Crossref Full Text | Google Scholar

78. Watson KH, Ciriegio AE, Pfalzer AC, Snow ALB, Diehl S, McDonell KE, et al. Psychiatric symptoms among adolescents and young adults with or without the huntingtin gene expansion. JNP. (2025) 37(3):230–7. doi: 10.1176/appi.neuropsych.20240104

PubMed Abstract | Crossref Full Text | Google Scholar

79. Abdollah Zadegan S, Coco HM, Reddy KS, Anderson KM, Teixeira AL, and Stimming EF. Frequency and pathophysiology of apathy in huntington disease: A systematic review and meta-analysis. JNP. (2023) 35:121–32. doi: 10.1176/appi.neuropsych.20220033

PubMed Abstract | Crossref Full Text | Google Scholar

80. Kempnich CL, Andrews SC, Fisher F, Wong D, Georgiou-Karistianis N, and Stout JC. Emotion recognition correlates with social-neuropsychiatric dysfunction in huntington’s disease. J Int Neuropsychol Soc. (2018) 24:417–23. Available online at: https://www.cambridge.org/core/journals/journal-of-the-international-neuropsychological-society/article/abs/emotion-recognition-correlates-with-socialneuropsychiatric-dysfunction-in-huntingtons-disease/8FD44106824FCD7CB93A254626D3868F (Accessed October 31, 2022).

PubMed Abstract | Google Scholar

81. Hendel RK, Hellem MNN, Hjermind LE, Nielsen JE, and Vogel A. On the association between apathy and deficits of social cognition and executive functions in Huntington’s disease. J Int Neuropsychol Soc. (2022) 29(4):369–76. doi: 10.1017/S1355617722000364

PubMed Abstract | Crossref Full Text | Google Scholar

82. Osborne-Crowley K, Andrews SC, Labuschagne I, Nair A, Scahill R, Craufurd D, et al. Apathy associated with impaired recognition of happy facial expressions in huntington’s disease. J Int Neuropsychol Soc. (2019) 25:453–61. doi: 10.1017/S1355617718001224

PubMed Abstract | Crossref Full Text | Google Scholar

83. Eddy CM and Rickards H. Social cognition and quality of life in Huntington’s disease. Front Psychiatry. (2022) 13:963457. doi: 10.3389/fpsyt.2022.963457

PubMed Abstract | Crossref Full Text | Google Scholar

84. Fritz NE, Boileau NR, Stout JC, Ready R, Perlmutter JS, Paulsen JS, et al. Relationships among apathy, health-related quality of life, and function in huntington’s disease. JNP. (2018) 30:194–201. doi: 10.1176/appi.neuropsych.17080173

PubMed Abstract | Crossref Full Text | Google Scholar

85. Connell J, O’Cathain A, and Brazier J. Measuring quality of life in mental health: Are we asking the right questions? Soc Sci Med. (2014) 120:12–20. Available online at: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4224500/.

PubMed Abstract | Google Scholar

86. Steinmetz JD, Seeher KM, Schiess N, Nichols E, Cao B, Servili C, et al. Global, regional, and national burden of disorders affecting the nervous system, 1990–2021: a systematic analysis for the Global Burden of Disease Study 2021. Lancet Neurol. (2024) 23:344–81. Available online at: https://www.thelancet.com/journals/laneur/article/PIIS1474-4422(24)00038-3/fulltext (Accessed February 06, 2025).

PubMed Abstract | Google Scholar

87. Douglas-Cowie E, Campbell N, Cowie R, and Roach P. Emotional speech: Towards a new generation of databases. Speech Communication. (2003) 40:33–60. Available online at: https://www.sciencedirect.com/science/article/pii/S0167639302000705 (Accessed December 07, 2023).

Google Scholar

88. Dhamyal H, Memon SA, Raj B, and Singh R. The phonetic bases of vocal expressed emotion: natural versus acted. arXiv preprint. (2020). Available online at: http://arxiv.org/abs/1911.05733 (Accessed January 10, 2025).

Google Scholar

89. Haamer RE, Rusadze E, Lüsi I, Ahmed T, Escalera S, Anbarjafari G, et al. Review on emotion recognition databases. In: Human-Robot Interaction - Theory and Application. IntechOpen (2017). Available online at: https://www.intechopen.com/chapters/58386 (Accessed January 11, 2025).

Google Scholar

90. Hantke S, Sagha H, Cummins N, and Schuller B. Emotional speech of mentally and physically disabled individuals: introducing the emotAsS database and first findings. In: Interspeech 2017. ISCA (2017). p. 3137–41. Available online at: https://www.isca-archive.org/interspeech_2017/hantke17_interspeech.html.

Google Scholar

91. Alhinti L, Cunningham S, and Christensen H. The Dysarthric Expressed Emotional Database (DEED): An audio-visual database in British English. PloS One. (2023) 18:e0287971. doi: 10.1371/journal.pone.0287971

PubMed Abstract | Crossref Full Text | Google Scholar

92. Kim D, Yi B, and Won Y. Speech emotion recognition in people at high risk of dementia. Dement Neurocogn Disord. (2024) 23:146–60. doi: 10.12779/dnd.2024.23.3.146

PubMed Abstract | Crossref Full Text | Google Scholar

93. Scherer KR. Vocal markers of emotion: Comparing induction and acting elicitation. Comput Speech Lang. (2013) 27:40–58. Available online at: https://www.sciencedirect.com/science/article/pii/S0885230811000647 (Accessed January 09, 2025).

Google Scholar

94. Riad R, Titeux H, Lemoine L, Montillot J, Bagnou JH, Cao XN, et al. Vocal markers from sustained phonation in Huntington’s Disease. arXiv preprint. (2020). Available online at: http://arxiv.org/abs/2006.05365 (Accessed January 02, 2024).

Google Scholar

95. Chenain L, Riad R, Fraisse N, Jubin C, Morgado G, Youssov K, et al. Graph methods to infer spatial disturbances: Application to Huntington’s Disease’s speech. Cortex. (2024) 176:144–60. Available online at: https://www.sciencedirect.com/science/article/pii/S0010945224001266 (Accessed June 11, 2024).

PubMed Abstract | Google Scholar

96. Kouba T, Frank W, Tykalova T, Mühlbäck A, Klempíř J, Lindenberg K, et al. Neurodiem. In: Automated speech analysis demonstrates utility in Huntington’s disease (2018). Available online at: https://www.neurodiem.fr/news/automated-speech-analysis-demonstrates-utility-in-huntingtons-disease-4gNu0Y7szgeoKLYQraL4dV?locale=en-US (Accessed April 24, 2023).

Google Scholar

97. Rusz J, Klempíř J, Baborová E, Tykalová T, Majerová V, Cmejla R, et al. Objective acoustic quantification of phonatory dysfunction in Huntington’s disease. PloS One. (2013) 8(6):e65881. doi: 10.1371/journal.pone.0065881

PubMed Abstract | Crossref Full Text | Google Scholar

98. Vogel AP, Shirbin C, Churchyard AJ, and Stout JC. Speech acoustic markers of early stage and prodromal Huntington’s disease: A marker of disease onset? Neuropsychologia. (2012) 50:3273–8. Available online at: https://www.sciencedirect.com/science/article/pii/S0028393212003831 (Accessed April 24, 2023).

PubMed Abstract | Google Scholar

99. Nunes AS, Pawlik M, Mishra RK, Waddell E, Coffey M, Tarolli CG, et al. Digital assessment of speech in Huntington disease. Front Neurol. (2024) 15:1310548. doi: 10.3389/fneur.2024.1310548

PubMed Abstract | Crossref Full Text | Google Scholar

100. Cummins N, Scherer S, Krajewski J, Schnieder S, Epps J, and Quatieri TF. A review of depression and suicide risk assessment using speech analysis. Speech Communication. (2015) 71:10–49. Available online at: https://www.sciencedirect.com/science/article/pii/S0167639315000369 (Accessed July 16, 2025)

Google Scholar

101. Laricchiuta D, Termine A, Fabrizio C, Passarello N, Greco F, Piras F, et al. Only words count; the rest is mere chattering: A cross-disciplinary approach to the verbal expression of emotional experience. Behav Sci. (2022) 12:292. Available online at: https://www.mdpi.com/2076-328X/12/8/292 (Accessed July 19, 2025).

PubMed Abstract | Google Scholar

102. Schuller BW. Speech emotion recognition: two decades in a nutshell, benchmarks, and ongoing trends. Commun ACM. (2018) 61:90–9. doi: 10.1145/3129340

Crossref Full Text | Google Scholar

103. Craufurd D. Problem Behaviours Assessment HD - Short Version (PBA-s) NINDS CDE Notice of Copyright (2024). Available online at: https://www.commondataelements.ninds.nih.gov/report-viewer/23770/Problem%20Behaviours%20Assessment%20HD%20-%20Short%20Version%20(PBA-s) (Accessed March 05, 2025).

Google Scholar

104. Frich JC, Rae D, Roxburgh R, Miedzybrodzka ZH, Edmondson M, Pope EB, et al. Health care delivery practices in huntington’s disease specialty clinics: an international survey. J Huntington’s Dis. (2016) 5:207–13. Available online at: https://content.iospress.com/articles/journal-of-huntingtons-disease/jhd160192 (Accessed July 24, 2023).

PubMed Abstract | Google Scholar

105. Skirton H, Williams JK, Jackson Barnette J, and Paulsen JS. Huntington disease: families’ experiences of healthcare services. J Adv Nurs. (2010) 66:500–10. doi: 10.1111/j.1365-2648.2009.05217.x

PubMed Abstract | Crossref Full Text | Google Scholar

106. van Lonkhuizen PJC, Frank W, Heemskerk AW, van Duijn E, de Bot ST, Mühlbäck A, et al. Quality of life, health-related quality of life, and associated factors in Huntington’s disease: a systematic review. J Neurol. (2023) 270:2416–37. doi: 10.1007/s00415-022-11551-8

PubMed Abstract | Crossref Full Text | Google Scholar

107. van Duijn E, Craufurd D, Hubers AAM, Giltay EJ, Bonelli R, Rickards H, et al. Neuropsychiatric symptoms in a European Huntington’s disease cohort (REGISTRY). J Neurol Neurosurg Psychiatry. (2014) 85:1411–8. Available online at: https://jnnp.bmj.com/content/85/12/1411 (Accessed February 06, 2025).

PubMed Abstract | Google Scholar

108. Ross CA and Tabrizi SJ. Huntington’s disease: from molecular pathogenesis to clinical treatment. Lancet Neurol. (2011) 10:83–98. Available online at: https://www.thelancet.com/journals/laneur/article/PIIS1474-4422(10)70245-3/abstract (Accessed February 02, 2025).

Google Scholar

Keywords: Huntington’s disease, emotional expression, psychiatric symptoms, speech analysis, remote monitoring

Citation: Chenain L, Fabre A, Titeux H, Morgado G, Youssov K, Clavel C and Bachoud-Lévi A-C (2025) Emotional speech markers of psychiatric disturbance in Huntington’s disease. Front. Psychiatry 16:1633492. doi: 10.3389/fpsyt.2025.1633492

Received: 22 May 2025; Accepted: 24 July 2025;
Published: 12 August 2025.

Edited by:

Perman Gochyyev, University of California, Berkeley, United States

Reviewed by:

Daniela Laricchiuta, University of Perugia, Italy
Yaneth Rodriguez-Agudelo, Manuel Velasco Suárez Instituto Nacional de Neurología y Neurocirugía, Mexico
Francisco Paz, Manuel Velasco Suárez Instituto Nacional de Neurología y Neurocirugía, Mexico

Copyright © 2025 Chenain, Fabre, Titeux, Morgado, Youssov, Clavel and Bachoud-Lévi. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Lucie Chenain, bHVjaWUuY2hlbmFpbkBlbnMucHNsLmV1

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.