BRIEF RESEARCH REPORT article

Front. Commun., 12 March 2024

Sec. Language Communication

Volume 9 - 2024 | https://doi.org/10.3389/fcomm.2024.1366597

You better listen to yourself: studying metacognitive efficiency in emotion recognition by voice

  • School of Occupational Therapy, Faculty of Medicine, Hebrew University of Jerusalem, Jerusalem, Israel

Abstract

Purpose:

Metacognition, or the ability to monitor the performance of oneself, is known for its fundamental importance for human behavior adjustments. However, studies of metacognition in social behaviors focused on emotion recognition are relatively scarce. In the current study, we aimed to examine the effectiveness of metacognition, measured by self-rated confidence in voice emotion recognition tasks within healthy individuals.

Methods:

We collected 180 audio-recorded lexical sentences portraying discrete emotions: anger, happiness, sadness, fear, surprise, and neutrality expressions. Upon listening to voice stimuli, participants (N = 100; 50 females, 50 males) completed the perception task of recognition of emotion. After each trial, a confidence rating (CR) was assigned.

Results:

A series of one-tailed t-tests showed that the differences in mean CRs assigned to correct and incorrect performances were significant for all emotions and neutral expression.

Conclusions:

Our preliminary results demonstrate efficiency in metacognition of emotion recognition by voice. Theoretically, our results support the difference between accuracy in metacognition, measured by CR, and efficiency in metacognition, as it specified the CR between correct and incorrect performance. To gain better insights into practical issues, further studies are needed to examine whether and what are the differences between accuracy and efficiency in metacognition, as part of social communication.

Introduction

Metacognition refers to the capacity to reflect on, monitor, and control cognitive processes, such as decision-making, memory, and perception (Flavell, 1979). Efficient metacognition commonly represents better self-assessment for one ability. While measuring metacognition ability may be achieved by assigning self-report confidence rating in a given task, efficiency in metacognition is commonly achieved by studying the difference in CR between correct and incorrect performance (Polyanskaya, 2023). Over decades, researchers have been motivated to learn the extensive role metacognition plays in executing adaptive human behaviors, mainly focusing on education and learning aspects (Baer and Kidd, 2022), with relatively less interest in emotional perception (Laukka et al., 2021; Rahnev, 2023). However, studying efficiency in metacognition is still scarce. Given the fundamental role metacognition took in emotion perception as a perceptional process, a better understanding of measuring efficiency in metacognition of emotion recognition is much needed. To fill this gap in the literature, in this preliminary study we aimed to investigate metacognitive efficiency in emotion recognition focusing on voice communication, across different emotions category.

Metacognitive efficiency

Efficient metacognition generally reflects a better introspection of cognitive process. Specifically, effective metacognition is suggested to promote effectiveness in noticing relevant cues and to seek additional information when necessary (Drigas and Mitsea, 2020; Brus et al., 2021). Therefore, metacognition is considered foundational to flexible and adaptive behavior in a range of settings (Baer and Kidd, 2022; Katyal and Fleming, 2024). Furthermore, promoting efficient metacognition by guided intervention had a positive impact on discrimination, working memory, and spatial tasks (Carpenter et al., 2019; McWilliams et al., 2023). A widely used method to study metacognition is to have observers do a challenging task followed by an explicit confidence judgment regarding their task performance (McWilliams et al., 2023). This retrospective CR is further used as a measure of monitoring effectiveness, when a larger difference in CRs assigned to correct and incorrect performances reflects the ability to estimate the likelihood of the observer making an error (Rahnev, 2021; Polyanskaya, 2023). Based on CR measurement, metacognitive efficiency is sometimes considered domain-specific, meaning that one's ability to effectively evaluate their correctness in performance is specific to the object under scrutiny. Hence, individuals can also have high awareness of their performance in one cognitive domain while having limited awareness of their performance in another (Rouault et al., 2018). As such, CRs for correct performance were lower in the perceptual task (i.e., visuospatial task) than in the memory task (Pennington et al., 2021). However, less studies posited a domain-general metacognitive competence based on audio, visual and audio visual modalities (Ais et al., 2016; Carpenter et al., 2019). Together, inconsistency in findings highlights the importance of independent metacognitive efficiency study for a range of domains and skills. Yet, most research remains focused on general knowledge and learning domains, with less interest in social domains (Saoud and Al-Marzouqi, 2020).

Metacognitive efficiency in emotion recognition

As metacognition contributes to accurate performance in other perception tasks, few researchers sought to gain further insights into its potential contribution to accurate perception of others' emotions (Jeckeln et al., 2022; Michel, 2023). Emotion perception competence refers to the ability to perceive and interpret (i.e., decode) relevant social–emotional cues expressed by others (i.e., to encode; Bänziger and Scherer, 2007). Accurate perception of others' feeling and thinking is a key skill associated with achieving successful interpersonal relationships (Mitsea et al., 2022). In the visual channel, for example, authors revealed that higher CR was associated with higher accuracy in the facial emotional recognition task, underscoring the fundamental role of CR in the adjustments of social behavior (Bègue et al., 2019; Garcia-Cordero et al., 2021; Folz et al., 2022; Katyal and Fleming, 2024), whereas metacognitive dysfunction has been implicated as a key source of maladaptive behavior regarding emotional perception (Arslanova et al., 2023).

When it comes to voice communication, studying efficiency in metacognition of emotional perception is relatively small in amount. Voice expressions can be either verbal (e.g., words and sentences) or nonverbal (e.g., screaming, laughing), where all have proven effective in conveying emotions (Eyben et al., 2015). In general, voice comprises unique audible features called prosody. Prosody is an important feature of language, mainly represented by intonation (i.e., how low or high a voice is perceived), loudness (i.e., how loud a sound is perceived), and tempo (i.e., how quickly an utterance is produced; Mitchell et al., 2003). Prosody also conveys important information about a speaker's sex, age, attitude, and mental state (Schwartz and Pell, 2012). Lausen and Hammerschmidt (2020) conducted one of the few studies examining CR's role in vocal emotion perception. They found that participants' confidence judgments increased with increasing correctness in emotion recognition tasks that included verbal and nonverbal stimuli. As a result, they speculated that specific characteristics of voiced stimuli affect CRs. They also found that meaning played a significant role in performance confidence as a higher confidence rating was assigned to stimuli containing lexical-semantic content (e.g., nouns and lexical sentences) rather than stimuli without meaning (e.g., pseudonouns and pseudosentences). More recently, Laukka et al. (2021) extended the concept of CRs to include multimodal analysis. On the basis of visual, audio, and visual–audio channels, the researchers also found a strong correlation between emotional recognition accuracies and CRs, in each channel (Laukka et al., 2021).

While the above studies demonstrate the importance of metacognitive accuracy, based on CR measurement, as far as we know, no study focused on metacognitive efficiency. In the current study, we deal for the first time with efficiency in metacognition concerning voice emotion recognition. Specifically, we asked whether correct performance in emotion recognition test would yield significant higher CR, compared to CR assigned to incorrect performance? As we intend to focus on voice communication, we adhere to previous suggestions that gender roles should be studied within a given societal context, particularly language (Keshtiari and Kuhlmann, 2016; Hall et al., 2021). As a result, we used an emotional target originally designed in Hebrew (Sinvani and Sapir, 2022) in the form of repeated lexical sentences, reflecting discrete emotions (anger, sadness, happiness, fear, and surprise) as well as neutral expression. In this way, we intended to focus on emotional prosodic cues separately from semantic cues. Earlier studies have discussed such methodologies (Paulmann and Kotz, 2008; Sinvani and Sapir, 2022), with particular recommendations in the diagnostic analysis of nonverbal accuracy 2 (DANVA-2), emphasizing the independent role of prosodic cues in emotion recognition (Nowicki and Duke, 2008). To answer our question, we used performance-based tests for emotion recognition by voice, followed by CR reported trial by trial. For the purpose of metacognitive efficiency detection, we set the mean CR scores for correct to incorrect performances. We hypothesis to find significant difference between CR assigned to correct and incorrect performance, where higher CR would be assigned to the correct performance. By hypothesizing that, we actually expect to illustrate here efficient in metacognition across all tested emotions, perceived by voice.

Materials and methods

Recording stage

Participants

A sample of recorders (i.e., encoders) comprised of 30 participants (16 men; 14 women). All encoders were actors (15 professional; 15 non-professional), native Hebrew speakers and were either undergraduate or graduate students at Haifa University, 20–35 years of age (M = 25.07, SD = 2.29). They had normal hearing, which was verified by a standard audiological assessment for pure tones from 0.5 to 8 kHz (Koerner and Zhang, 2018). None had any history of speech, language, or hearing disorders or psychiatric illnesses.

Procedure

Before the experiment commenced, participants were given a description of the experimental procedure and their tasks. Participants then signed a consent form and completed a short demographic questionnaire concerning age and gender. The study procedure included a brief hearing test and a recording stage. All procedures were conducted in the same acoustic room in the Interdisciplinary Clinics Center at Haifa University. Stimuli were repeated lexical sentence in Hebrew, previously reported in Sinvani and Sapir (2022), meaning “Oh really? I can't believe it!” [/Ma/Be/e/met/A/ni/lo//Ma/a/mi/na/]. Here, the same experimenter (first author) instructed the encoders to vocally enact the repeated target sentence in five discrete emotions: anger, happiness, sadness, fear, and surprise. The elicitation of audio samples followed a design of 30 actors × five emotions × 1 sentence, resulting in 150 emotional utterances. In addition, each of the 30 actors uttered the sentence in a neutral, non-emotional fashion, yielding additional 30 neutral stimuli. The speech recordings were obtained in a sound-proof booth using a head-mounted condenser microphone (AKG C410) positioned 10 cm and 45–50° from the left oral angle. The signal was pre-amplified, low pass-filtered at 9.8 kHz and digitized to a computer hard disk at a sampling rate of 22 kHz (Stipancic and Tjaden, 2022), using Cool Edit Pro version 2.0. All audio clips were auditory scanned to ensure adequate amplitude and no missing data (i.e., speakers produced all utterances). The full set of audio clips is available from the authors upon request.

Perceptual tests: emotion recognition task and SRC

Participants

The sample of listeners (i.e., decoders) comprised of 100 participants (50 men; 50 women). All decoders were native Hebrew speakers and were either undergraduate or graduate students at Haifa University, 20–35 years of age (M = 24.37, SD = 2.28). All had normal hearing, which was verified by a standard audiological assessment for pure tones from 0.5 to 8 kHz (Koerner and Zhang, 2018). None had any history of speech, language, or hearing disorders or psychiatric illnesses.

Procedure

Before the experiment commenced, participants were given a description of the experimental procedure and their tasks. Participants then signed a consent form and completed a short demographic questionnaire concerning age and gender. The study procedure included a brief hearing test and an emotion recognition task, followed by retrospective SRC regarding emotion recognition accuracy. All procedures were conducted in the same acoustic room in the Interdisciplinary Clinics Center at Haifa University.

Ten decoders (five men; five women) were invited to each experimental session, which lasted ~10 min. Upon arrival, the experimenter informed the participants about the aim and procedure of the study. Decoders were informed that each stimulus would be presented only once. For each experimental session, 10 decoders judged the same 18 target stimuli, created by three encoders. In each session, targets were randomized in order. In addition, decoders were blinded to the ratings of the remaining decoders, as each decoder sat alone in the acoustic room with headphones for the entire procedure. Stimuli were presented to the participants binaurally via Sennheiser HD 448 headphones plugged in the tower box of a Dell OptiPlex 780 SFF desktop PC computer. In total, the perceptual test generated 1,800 results for ERA and 1,800 results for SRC.

Measures

Voice emotion recognition task

Emotion recognition accuracy was assessed using a performance-based test in a categorical manner. Upon listening to each auditory stimulus, decoders marked one of six categories: happiness, anger, sadness, fear, surprise, or neutrality. As innate and universal emotions, these five categories are frequently referred to as basic emotions (Ekman, 1992; Siedlecka and Denson, 2019). Moreover, we included neutral expressions, as commonly done in the literature on voice emotional recognition (Lausen and Schacht, 2018; Lin et al., 2021). In accordance with the DANVA-2 paradigm, we implemented a repeat lexical sentence designed in Hebrew, meaning “Oh really?” “I can't believe it!” [/Ma/Be/e/met/A/ni/lo//Ma/a/mi/na/]. This target was used in a previous study conducted in our lab (Sinvani and Sapir, 2022). In a preliminary test, we confirmed that identification percentages were significantly greater than chance level, in each of the tested category (Figure 1).

Figure 1

Retrospective SRC

CR was measured retrospectively by self-reported own confidence in a given response (in the voice emotion recognition task). Based on a Likert-type scale from 0 (not at all confident) to 6 (extremely confident), decoders rated how confident they were in their emotion category judgments. A similar 7-point scale was used in a previous study (Bègue et al., 2019) to report SRC in facial expression recognition.

Ethics

The IRB Ethics Committee of the University of Haifa, Israel approved the study. Participants were recruited using convenience sampling through social media advertisements and word of mouth. Professional and non-professional actors consented to the anonymous use of their shortened recordings in our study. Actors in the recording stage and participants in the perceptual test were paid 20 NIS for their participation.

Data analysis

Statistical analysis was conducted using IBM SPSS (Statistical Package for the Social Sciences) version 28.0. For all analyses, p < 0.05 was considered statistically significant. Independent-samples t-tests were conducted to examine the differences in CRs by the performance (correct/incorrect) across all emotions. The significance is reported controlling for false discovery rate for multiple comparisons (Benjamini and Hochberg, 1995).

Results

We examined metacognitive efficiency in the magnitude of the difference in CRs assigned to correct (i.e., where target emotion fits the response emotion) and incorrect (i.e., where target emotion was different from response emotion) performances for all decoders.

Figure 1 shows that correct performances were assigned higher confidence than incorrect performances for all emotions. A series of one-tailed t-tests showed that the differences in mean CRs assigned to correct and incorrect performances were significant for all emotions and neutral expression (Table 1).

Table 1

EmotionPerformance
IncorrectCorrecttOne-tailed adjustedp-valuea
NM(SD)NM(SD)
Afraid1964.43 (1.25)1044.67 (1.14)−1.6760.048
Angry1764.29 (1.38)1244.65 (1.12)−2.4540.014
Happy1254.19 (1.45)1755.07 (1.10)−5.731< 0.001
Neutral1144.24 (1.34)1864.52 (1.29)−1.7790.046
Sad773.74 (1.47)2234.90 (1.13)−6.310< 0.001
Surprised1484.46 (1.25)1524.72 (1.30)−1.7940.046

Means and SD of CRs assigned to correct and incorrect performance by emotion.

aBenjamini-Hochberg adjustment.

Discussion

In the current preliminary study, we investigated the metacognition efficiency of an emotion recognition task performance, conveyed by voice. To this end, we analyzed differences in CRs between correct and incorrect performance in the emotion recognition task, based on identical vocal lexical sentences portrayed into emotional prosodies of anger, happiness, sadness, surprise and fear, and neutral expression. Following the literature dealing with the global concept of metacognitive efficiency, we expected to demonstrate a greater difference between CRs assigned to correct and incorrect performances, thus illustrating a convenient representation of metacognitive efficacy. As far as we know, this is the first study conducted to examine efficiency in metacognition skills concerning emotional recognition by voice.

As predicted, our results indicate a consistent difference in CRs between correct and incorrect performance, where higher CRs were significantly assigned for correct performances. In accordance with previous studies, our findings support the general established finding for metacognition efficiency in diverse cognitive tasks (Ais et al., 2016; Carpenter et al., 2019). To our knowledge, only two studies have been conducted to examine metacognitive ability based on retrospective confidence rating for vocal emotion recognition (Lausen and Hammerschmidt, 2020; Laukka et al., 2021). Although these studies based their examination for metacognitive skill on the measure of CR in a given response, representing metacognitive accuracy, they ignore the difference in CR between correct to incorrect performance, comprising the metacognitive efficiency.

Following extensive study focusing on brain mechanism underlying correctness in emotion recognition tasks, one of the worth noting question concerns the neural selectivity is if it is determined by the emotion presented by encoder or by the emotion perceived by a decoder (Wang et al., 2014). Based on face stimuli, Wang et al. (2014) concluded that neurons in the human amygdala encode the perceived judgment of emotions rather than the stimulus features themselves. Recently, Zhang et al. (2022) further showed that responses of anterior insula and amygdala were driven by the perceived vocal emotions. However, they revealed that responses of Heschl's gyrus and posterior insula (i.e., both involved in auditory perception) were determined by the presented emotion. Together, it was speculated that extensive connectivity of the amygdala with other brain regions allows it to modulate a few functions essential in adaptive and socio-emotional behaviors (Whitehead, 2021; Atzil et al., 2023). Given the fundamental role metacognition (i.e., confidence) plays in adaptive and socio-emotional behaviors, our results highlight the further need for incorporating neuroscience study into neural mechanism underlying self-confidence to distinguish correct and incorrect performance in emotion recognition task. This study has theoretical and practical significance. From the theoretic point of view, our results confirm earlier studies concerning various cognitive processes by establishing that individuals are capable of efficiently assessing their performance in voice emotion recognition (Cho, 2017; Skewes et al., 2021; Polyanskaya, 2023). Moreover, given the significant association that was indicated across all emotions and neutral expressions, we speculated metacognitive efficiency of being individual “fingerprint”, rather than domain-specific (Wang et al., 2014; Ais et al., 2016; Zhang et al., 2022). Additional studies should incorporate neuroscience methodologies in diverse channels of emotion communication to specify the mechanism of efficiently introspecting vocal emotion recognition to gain support for this claim (Wang et al., 2014; Whitehead, 2021; Zhang et al., 2022; Atzil et al., 2023).

From the practice point of view, in light of previous results showing the increase in performance found among students after they participate in guided intervention for metacognitive awareness (Nunaki et al., 2019), our findings should be taken one step further regarding designing interventions focused on emotion recognition through voice. Unlike vocal channel which deserved much less attention, several studies demonstrated the effectiveness of practicing facial emotion recognition skill as beneficial to social interactions (Chaidi and Drigas, 2020; Dantas and do Nascimento, 2022a,b). Importantly, those works highlight the flexibility of facial emotion recognition skill which in our opinion justify comprehensive research in the voice channel. In addition, our findings may consider valuable by diagnosis means, as extensive research dealt with metacognitive maladaptive in diverse subclinical and clinical populations (Hoven et al., 2019).

Although our study offers a valuable contribution to the current understanding of metacognition in social communication of emotions, some limitations should be noted. First, to investigate emotional recognition, we used a lexical sentence as a repeated item. Therefore, one can argue that prosodic cues are more important than semantics. However, lexical sentences can still affect perception and confidence by contributing meaning (Lausen and Hammerschmidt, 2020). Therefore, researchers should consider various types of stimuli for a better generalization of findings regarding efficiency in metacognition of emotion recognition by voice. Second, it is important to note that our investigation was blinded to acoustical parameters, which have been documented as contributing to the CR discussed elsewhere (Lausen and Hammerschmidt, 2020). Spectral parameters and prosodic contours (Shaqra et al., 2019) would be valuable tools for future research, allowing researchers to detect physiological changes in voice and emotional expressions.

Conclusions

Our study offers a preliminary investigation of metacognitive efficiency in emotion recognition. As far as we know this is the first study to focus on CR for measuring efficacy in emotion recognition task, and not just accuracy in metacognition. Our findings highlight the need for better understand in how individuals utilize metacognitive efficiency in their social interactions. We recommend further studies to investigate efficiency by comparing different groups and different channels of emotion communication. As a result of futural studies we will be able to understand the distinct contribution of efficiency over accuracy in social emotional communication as a fundamental for one's mental health.

Statements

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by the IRB Ethics Committee of the University of Haifa, Israel. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

R-TS: Conceptualization, Data curation, Formal analysis, Methodology, Writing—original draft, Writing—review & editing. HF-G: Conceptualization, Data curation, Formal analysis, Writing – original draft.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

In memoriam

The authors feel honored to dedicate this work in memory of Professor Shimon Sapir and his inspiring leadership and vigorous role in advancing acoustic speech analysis.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1

    AisJ.ZylberbergA.BarttfeldP.SigmanM. (2016). Individual consistency in the accuracy and distribution of confidence judgments. Cognition. 146, 377386. 10.1016/j.cognition.2015.10.006

  • 2

    ArslanovaI.MeletakiV.Calvo-MerinoB.ForsterB. (2023). Perception of facial expressions involves emotion specific somatosensory cortex activations which are shaped by alexithymia. Cortex. 167, 223234. 10.1016/j.cortex.2023.06.010

  • 3

    AtzilS.SatputeA. B.ZhangJ.ParrishM. H.ShablackH.MacCormackJ. K.et al. (2023). The impact of sociality and affective valence on brain activation: a meta-analysis. Neuroimage. 268, 119879. 10.1016/j.neuroimage.2023.119879

  • 4

    BaerC.KiddC. (2022). Learning with certainty in childhood. Trends Cogn Sci. 26, 887896. 10.1016/j.tics.2022.07.010

  • 5

    BänzigerT.SchererK. R. (2007). Using actor portrayals to systematically study multimodal emotion expression: The GEMEP corpus. In: Affective Computing and Intelligent Interaction: Second International Conference, ACII 2007. Lisbon: Springer.

  • 6

    BègueI.VaessenM.HofmeisterJ.PereiraM.SchwartzS.VuilleumierP.et al. (2019). Confidence of emotion expression recognition recruits brain regions outside the face perception network. Soc. Cognit. Affect. Neurosci. 14, 8195. 10.1093/scan/nsy102

  • 7

    BenjaminiY.HochbergY. (1995). Controlling the false discovery rate: a practical and powerful approach to multiple testing. J. R. Stat. Soc. B. 57, 289300. 10.1111/j.2517-6161.1995.tb02031.x

  • 8

    BrusJ.AebersoldH.GrueschowM.PolaniaR. (2021). Sources of confidence in value-based choice. Nat Commun. 12, 7337. 10.1038/s41467-021-27618-5

  • 9

    CarpenterJ.ShermanM. T.KievitR. A.SethA. K.LauH.FlemingS. M.et al. (2019). Domain-general enhancements of metacognitive ability through adaptive training. J. Exp. Psychol. Gen. 148, 51. 10.1037/xge0000505

  • 10

    ChaidiI.DrigasA. (2020). Autism, Expression, and Understanding of Emotions: Literature Review. International Association of Online Engineering. Retrieved from: https://www.learntechlib.org/p/218023/ (accessed February 25, 2024).

  • 11

    ChoS-. Y. (2017). Explaining Gender Differences in Confidence and Overconfidence in Math. 10.2139/ssrn.2902717

  • 12

    DantasA.do NascimentoM. (2022b). Face emotions: improving emotional skills in individuals with autism. Mul. Tools Appl. 81, 2594725969. 10.1007/s11042-022-12810-6

  • 13

    DantasA. C.do NascimentoM. Z. (2022a). Recognition of emotions for people with autism: an approach to improve skills. Int J Comp Games Technol. 2022, 121. 10.1155/2022/6738068

  • 14

    DrigasA.MitseaE. (2020). The 8 pillars of metacognition. Int. J. Emerg. Technol. Learn. 15, 162178. 10.3991/ijet.v15i21.14907

  • 15

    EkmanP. (1992). An argument for basic emotions. Cogn Emot. 6, 169200. 10.1080/02699939208411068

  • 16

    EybenF.SchererK. R.SchullerB. W.SundbergJ.Andr,éE.BussoC.et al. (2015). The Geneva minimalistic acoustic parameter set (GeMAPS) for voice research and affective computing. IEEE Trans. Affect. Comput. 7, 190202. 10.1109/TAFFC.2015.2457417

  • 17

    FlavellJ. H. (1979). Metacognition and cognitive monitoring: a new area of cognitive–developmental inquiry. Am. Psychol. 34, 906. 10.1037/0003-066X.34.10.906

  • 18

    FolzJ.AkdağRNikolićMvan SteenbergenH.KretM. E. (2022). Facial mimicry and metacognitive judgments in emotion recognition-modulated by social anxiety and autistic traits? Available online at: https://psyarxivcom/e7w6k/ (accessed 26, 2022).

  • 19

    Garcia-CorderoI.MigeotJ.FittipaldiS.AquinoA.CampoC. G.GarcíaA.et al. (2021). Metacognition of emotion recognition across neurodegenerative diseases. Cortex.137, 93107. 10.1016/j.cortex.2020.12.023

  • 20

    HallK.BorbaR.HiramotoM. (2021). Language and gender. Int. Encycl. Linguis. Anthropol. 2021, 892912. 10.1002/9781118786093.iela0143

  • 21

    HovenM.LebretonM.EngelmannJ. B.DenysD.LuigjesJ.van HolstR. J. (2019). Abnormalities of confidence in psychiatry: an overview and future perspectives. Transl. Psychiatry9:268. 10.1038/s41398-019-0602-7

  • 22

    JeckelnG.MamassianP.O'tooleA. J. (2022). Confidence judgments are associated with face identification accuracy: findings from a confidence forced-choice task. Behav. Res. Methods. 2022, 110. 10.31234/osf.io/3dxus

  • 23

    KatyalS.FlemingS. M. (2024). The future of metacognition research: balancing construct breadth with measurement rigor. Cortex171, 223234. 10.1016/j.cortex.2023.11.002

  • 24

    KeshtiariN.KuhlmannM. (2016). The effects of culture and gender on the recognition of emotional speech: evidence from Persian speakers living in a collectivist society. Int. J. Soc. Cult. Lang. 4, 71. 10.13140/RG.2.1.1159.0001

  • 25

    KoernerT. K.ZhangY. (2018). Differential effects of hearing impairment and age on electrophysiological and behavioral measures of speech in noise. Hear. Res. 370, 130142. 10.1016/j.heares.2018.10.009

  • 26

    LaukkaP.BänzigerT.IsraelssonA.CortesD. S.TornbergC.SchererK. R.et al. (2021). Investigating individual differences in emotion recognition ability using the ERAM test. Acta. Psychol. 220, 103422. 10.1016/j.actpsy.2021.103422

  • 27

    LausenA.HammerschmidtK. (2020). Emotion recognition and confidence ratings predicted by vocal stimulus type and prosodic parameters. Hum. Soc. Sci. Commun. 7, 117. 10.1057/s41599-020-0499-z

  • 28

    LausenA.SchachtA. (2018). Gender differences in the recognition of vocal emotions. Front. Psychol. 9, 882. 10.3389/fpsyg.2018.00882

  • 29

    LinY.DingH.ZhangY. (2021). Gender differences in identifying facial, prosodic, and semantic emotions show category-and channel-specific effects mediated by encoder's gender. J. Speech Lang. Hear. Res. 64, 29412955. 10.1044/2021_JSLHR-20-00553

  • 30

    McWilliamsA.BibbyH.SteinbeisN.DavidA. S.FlemingS. M. (2023). Age-related decreases in global metacognition are independent of local metacognition and task performance. Cognition235, 105389. 10.1016/j.cognition.2023.105389

  • 31

    MichelC. (2023). Towards a new standard model of concepts? Abstract concepts and the embodied mind: Rethinking grounded cognition, by Guy Dove, Oxford University Press, Oxford, 2022, 280. pp., $74.00, ISBN 9780190061975. Abingdon-on-Thames: Taylor & Francis.

  • 32

    MitchellR. L.ElliottR.BarryM.CruttendenA.WoodruffP. W. (2003). The neural response to emotional prosody, as revealed by functional magnetic resonance imaging. Neuropsychologia41, 14101421. 10.1016/S0028-3932(03)00017-4

  • 33

    MitseaE.DrigasA.SkianisC. (2022). Breathing, attention & consciousness in sync: the role of breathing training, metacognition & virtual reality. Technium Soc Sci J. 29, 79. 10.47577/tssj.v29i1.6145

  • 34

    NowickiS.DukeM. (2008). Manual for the Receptive Tests of the Diagnostic Analysis of Nonverbal Accuracy 2 (DANVA2). Atlanta, GA: Emory University.

  • 35

    NunakiJ. H.DamopoliiI.KandowangkoN. Y.NusantariE. (2019). The effectiveness of inquiry-based learning to train the students' metacognitive skills based on gender differences. Int. J. Instruct.12, 505516. 10.29333/iji.2019.12232a

  • 36

    PaulmannS.KotzS. A. (2008). An ERP investigation on the temporal dynamics of emotional prosody and emotional semantics in pseudo-and lexical-sentence context. Brain Lang. 105, 5969. 10.1016/j.bandl.2007.11.005

  • 37

    PenningtonC.BallH.SwirskiM.NewsonM.CoulthardE. (2021). Metacognitive performance on memory and visuospatial tasks in functional cognitive disorder. Brain Sci. 11, 1368. 10.3390/brainsci11101368

  • 38

    PolyanskayaL. (2023). I. know that I know. But do I know that I do not know?Front. Psychol. 14, 1128200. 10.3389/fpsyg.2023.1128200

  • 39

    RahnevD. (2021). Visual metacognition: Measures, models, and neural correlates. Am. Psychol. 76, 1445. 10.1037/amp0000937

  • 40

    RahnevD. (2023). Measuring metacognition: a comprehensive assessment of current methods. 10.31234/osf.io/waz9h

  • 41

    RouaultM.SeowT.GillanC. M.FlemingS. M. (2018). Psychiatric symptom dimensions are associated with dissociable shifts in metacognition but not task performance. Biol. Psychiatry. 84, 443451. 10.1016/j.biopsych.2017.12.017

  • 42

    SaoudL. S.Al-MarzouqiH. (2020). Metacognitive sedenion-valued neural network and its learning algorithm. IEEE Access. 8, 144823144838. 10.1109/ACCESS.2020.3014690

  • 43

    SchwartzR.PellM. D. (2012). Emotional speech processing at the intersection of prosody and semantics. PLoS ONE. 7, e47279. 10.1371/journal.pone.0047279

  • 44

    ShaqraF. A.DuwairiR.Al-AyyoubM. (2019). Recognizing emotion from speech based on age and gender using hierarchical models. Procedia. Comput. Sci. 151, 3744. 10.1016/j.procs.2019.04.009

  • 45

    SiedleckaE.DensonT. F. (2019). Experimental methods for inducing basic emotions: a qualitative review. Emot. Rev. 11, 8797. 10.1177/1754073917749016

  • 46

    SinvaniR.-T.SapirS. (2022). Sentence vs. word perception by young healthy females: toward a better understanding of emotion in spoken language. Front. Glob. Women's Health. 3, 829114. 10.3389/fgwh.2022.829114

  • 47

    SkewesJ.FrithC.OvergaardM. (2021). Awareness and confidence in perceptual decision-making. Brain Multiphy. 2, 100030. 10.1016/j.brain.2021.100030

  • 48

    StipancicK. L.TjadenK. (2022). Minimally detectable change of speech intelligibility in speakers with multiple sclerosis and Parkinson's disease. J. Speech Lang. Hear. Res. 65, 18581866. 10.1044/2022_JSLHR-21-00648

  • 49

    WangS.TudusciucO.MamelakA. N.RossI. B.AdolphsR.RutishauserU.et al. (2014). Neurons in the human amygdala selective for perceived emotion. Proc. Nat. Acad. Sci. 111, E3110E9. 10.1073/pnas.1323342111

  • 50

    WhiteheadJ. C. (2021). Investigating the Neural Correlates of Social and Emotional Processing Across Modalities. Montreal, QC: McGill University.

  • 51

    ZhangY.ZhouW.HuangJ.HongB.WangX. (2022). Neural correlates of perceived emotions in human insula and amygdala for auditory emotion recognition. Neuroimage260, 119502. 10.1016/j.neuroimage.2022.119502

Summary

Keywords

social interaction, language, voice communication, speech perception, self-confidence, prosody, metacognitive efficiency, metacognition

Citation

Sinvani R-T and Fogel-Grinvald H (2024) You better listen to yourself: studying metacognitive efficiency in emotion recognition by voice. Front. Commun. 9:1366597. doi: 10.3389/fcomm.2024.1366597

Received

06 January 2024

Accepted

19 February 2024

Published

12 March 2024

Volume

9 - 2024

Edited by

Martina Micai, National Institute of Health (ISS), Italy

Reviewed by

Adilmar Coelho Dantas, Federal University of Uberlandia, Brazil

Yang Zhang, Johns Hopkins University, United States

Updates

Copyright

*Correspondence: Rachel-Tzofia Sinvani

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics