SYSTEMATIC REVIEW article

Front. Psychol., 02 February 2023

Sec. Cognition

Volume 14 - 2023 | https://doi.org/10.3389/fpsyg.2023.991339

Metacognitive sensitivity and symptoms of mental disorder: A systematic review and meta-analysis

  • School of Psychology, University of Nottingham, Nottingham, United Kingdom

Article metrics

View details

3

Citations

5,1k

Views

1,3k

Downloads

Abstract

Introduction:

Metacognition is a term used to refer to cognition about cognitive processes. In this systematic review and meta-analysis, we reviewed studies that investigated the relationship between experimentally measured objective metacognitive sensitivity and diverse symptoms of mental disorder. In these studies, metacognitive sensitivity is operationalized as the correspondence between the accuracy of task performance and reported confidence therein.

Methods:

A literature search was conducted across four databases and studies were selected for review based on predefined eligibility criteria. Twenty studies were included in the review and separate meta-analyses were conducted for psychotic and non-psychotic categories of psychiatric symptoms.

Results:

A significant reduction (medium effect size) in metacognitive sensitivity was found in individuals with psychosis-related symptoms of mental disorder compared to healthy control groups, but no significant difference was found for individuals with non-psychotic symptoms. It should be noted though, that fewer studies were available for the latter group. Sub-group analysis found no evidence that the effect of metacognitive impairment depended on whether perceptual or non-perceptual experimental tasks were employed.

Discussion:

These findings are discussed in relation to other conceptualizations of metacognition and the role reduced metacognitive sensitivity may play in forms of mental disorder.

1. Introduction

Cognition can be described as the mental processes through which “sensory input is transformed, reduced, elaborated, stored, recovered, and used” (Neisser, 1967). Mental processes that are considered to be “cognition about cognitive phenomena” (Flavell, 1979, p. 906) rather than sensory input have been termed ‘metacognition’. Metacognition is usually distinguished from the operations of executive function, which also have other cognitive processes as their object, in that the former is considered to require conscious representation of the cognitive process in question (Heyes et al., 2020). Forms of metacognition have been differentiated based on the kinds of cognitive processes involved. The cognitive phenomena that are the object of metacognition, referred to as “first-order cognition,” can be assigned to different cognitive domains such as memory and perception, and are employed in monitoring and regulating interaction with stimuli. Cognition about these first-order processes, i.e., metacognition or “second-order cognition,” monitors and regulates the represented first-order processes. Performance in tasks requiring second-order cognition, referred to as “type-2” tasks (Clarke et al., 1959), has been found to be dissociable from that in tasks based only on first-order cognition, referred to as “type-1” tasks (e.g., Fleming et al., 2010; Rouault et al., 2018b). Where monitoring of first-order cognition produces a representation of the cognitive process that closely reflects its relation to environmental input, this can be considered high metacognitive sensitivity or accuracy. Whether the representation is of one’s own cognitive processes or those of others, accurate metacognition promotes adaptive behavior at an individual and interpersonal level (David et al., 2012) and has thus been an avenue for developing treatments of mental disorder (Moritz and Lysaker, 2018).

Metacognition can be considered at either the local or global level (Seow et al., 2021). The global level of metacognition pertains to how an individual monitors and regulates patterns of mental processes that can be considered general properties of themselves or of others across contexts, an example of which may be an individual’s perception of their ability to recognize previously encountered faces compared to their ability to recognize objects. On the other hand, local metacognition refers to an individual’s monitoring and regulation of their own mental activity where it is integral to the performance of a discrete task, such as the extent to which their recognition of a particular object is a reliable indication that they previously encountered that object. The two levels of metacognition are measured using different approaches. A profile of global metacognition is typically produced by asking individuals to report beliefs regarding their mental processes generally or conducting interviews to infer these attributions (e.g., Morrison et al., 2007; Sellers et al., 2017), whereas local metacognition is measured by comparing individuals’ first-order cognition, assessed objectively through behavioral tasks, to their self-reported perceptions of performance within these tasks.

Both local and global metacognition have been assessed alongside indicators of mental health in clinical (Davies et al., 2018) and non-clinical populations (Chan et al., 2015), as well as using transdiagnostic approaches (Rouault et al., 2018b). Research in the latter case has occurred as part of a movement toward measuring symptom dimensions to overcome problems arising from comorbidity and symptom variability within diagnostic categories (Gillan et al., 2016; Seow et al., 2021). Both local and global manifestations of metacognition have been found to be associated with symptoms of mental disorder (Cooper and Osman, 2007; Bliksted et al., 2017) as well as clinical insight, which depends on accurate assessment of cognitive functions where these are related to clinical diagnosis. For this reason, metacognition has been a target for change in treatment settings through various programs (Van Oosterhout et al., 2016; Moritz and Lysaker, 2018). These programmes, which promote adaptive ways of relating to mental experience, have been tailored more to global metacognition than local metacognition (Lysaker et al., 2018). Seow et al. (2021) however, argue that there is likely a bidirectional influence between the two levels of metacognition, given evidence that global metacognition is likely to influence how an individual monitors and regulates their task-specific cognition (Rouault and Fleming, 2020), while metacognition employed with respect to a particular task shapes a person’s global self-performance attributions (Rouault et al., 2019; Lee et al., 2021). Local metacognition (hereafter “metacognition” unless otherwise specified) measured experimentally combines behavioral measures from a standardized task and self-report measures relating to perceived performance. This operationalization of metacognition provides a standardized index of first-order cognitive processes for comparison with and evaluation of self-reported perceptions of cognition.

Metacognition is most often gauged by eliciting post-response statements of confidence in task performance. Using this experimental operationalization, metacognitive monitoring is the relationship between confidence and response accuracy, where adaptive metacognition is demonstrated through high confidence reports after accurate task performance and low confidence after inaccurate task performance. Metacognitive evaluation expressed as confidence with regards to response accuracy is typically considered to be influenced by two distinct properties (Maniscalco and Lau, 2012), which are bias and sensitivity. This review will focus on metacognitive sensitivity, or the extent to which an individual’s confidence discriminates their first-order accuracy. Some researchers have inferred metacognitive sensitivity simply through the strength of correlation between confidence and decision accuracy (Nelson et al., 1986), while others have used metrics derived from signal detection theory (SDT) to index “type-2” sensitivity (Galvin et al., 2003; Masson and Rotello, 2009) separately from response bias.

In tasks of first-order cognition, SDT is taken to measure response sensitivity, type-1 d’, independently from response bias where two external stimulus alternatives exist by comparing responses indicative of a stimulus when it is present, a participant’s “hits,” to such responses when the stimulus is absent, a participant’s “false alarms.” This analysis has been extended to measure metacognitive or “type-2” sensitivity by considering “hits” as responses of high confidence where accurate first-order performance is present in the form of type-1 hits or correct rejections, and “false alarms” where high confidence responses follow inaccurate first-order performance in the form of type-1 misses or false alarms. Type-1 and type-2 responses characterized according to SDT are described in Tables 1, 2, respectively.

TABLE 1

Signal-present responseSignal-absent response
Signal presentHit*Miss **
Signal absentFalse alarm **Correct rejection*

Type-1 SDT categories for signal responses as related to signal presence, * and ** denote the correct and incorrect type-1 responses.

TABLE 2

High confidenceLow confidence
Correct type-1 response (*)Type-2 hitType-2 miss
Incorrect type-1 response (**)Type-2 false alarmType-2 correct rejection

Type-2 SDT categories for confidence responses and their relation to type-1 accuracy.

A correct type-1 response that is judged as being given with high confidence would be a type-2 hit, whereas an incorrect type-1 response that is met with high confidence is a type-2 false alarm. See Fleming and Dolan (2012). * and ** denote the correct and incorrect type-1 responses.

As confidence responses, unlike type-1 responses, are not considered to fulfill SDT assumptions of Gaussian distribution (Galvin et al., 2003), researchers have implemented SDT analysis of type-2 responses by calculating Receiver Operating Characteristic (ROC) curves based on hit and false alarm rates for individual confidence criteria, where the area under a type-2 ROC curve can be interpreted as a measure of metacognitive sensitivity.

Type-2 ROC measures are however influenced by type-1 performance (Galvin et al., 2003; Evans and Azzopardi, 2007), as illustrated in Figure 1, which shows theoretical ROC curves for the same individual for different levels of type-1 d’. To avoid confounding metacognitive sensitivity with first-order accuracy, a model-based SDT metric has been developed more recently which takes type-1 performance into account for estimates of metacognitive sensitivity (Maniscalco and Lau, 2012).

FIGURE 1

FIGURE 1

The graph shows hypothetical type II ROC curves of a hypothetical observer with a given metacognitive capacity, exposed to different levels of difficulty in the primary task, implying different levels of type I sensitivity (d’). Empirical metacognitive sensitivity (type II) will be constrained by type I task difficulty - the same observer will show different levels of type II sensitivities, depending on the difficult in the primary task. This is why we only considered studies that had approximately equal type I performance levels for both participant groups. See also Fleming and Dolan (2012).

The influence of subjective task difficulty on measures of metacognitive sensitivity can be intuitively understood by considering the following example. In a visual detection task, a person would practically always be able to say whether they made an error in indicating the presence of a stimulus, i.e., that they have low confidence in their incorrect response, when that stimulus is highly visible to them in every “stimulus-present” trial, which would suggest perfect metacognitive sensitivity. However, the same individual would less reliably match their confidence to their accuracy in the same task with a less visible stimulus. This reduction in estimates of metacognitive sensitivity for difficult tasks is undesirable when attempting to assess an individual’s metacognitive sensitivity, which may be considered to be a stable property with respect to any task involving the same first order processes (Fleming et al., 2010). Where research attempts to assess metacognitive sensitivity in populations with different levels of psychiatric symptoms, which are likely to also differ in their first-order cognitive performance (Chan, 2016; Davies and Greenwood, 2020), it is particularly relevant to control for the latter either through experimental design or response analysis. A selection of the measures that have been used in estimating metacognitive sensitivity are summarized in Table 3.

TABLE 3

MeasureDescriptionGreater metacognitive sensitivity indicated by:Limitations
Low confident incorrect responsesNumber of incorrect responses with low confidenceHigher valueInfluenced by first-order performance and metacognitive bias
High confident incorrect responsesNumber of incorrect responses with high confidenceLower valueInfluenced by first-order performance and metacognitive bias
Goodman-Kruskall gamma coefficient
Goodman and Kruskal (1954)
Correlation between confidence and accuracyHigher valueInfluenced by first-order performance and metacognitive bias
Confidence gap Moritz et al. (2003)Mean confidence for correct responses - mean confidence for incorrect responseHigher valueInfluenced by first-order performance
Knowledge Corruption Index
Moritz et al. (2004)
Proportion of high confidence responses given for incorrect trials compared to total number of high confidence responsesLower valueInfluenced by first- order performance
AUROC2 Galvin et al. (2003)Area under the Receiver Operating Characteristic (ROC) curve for type-2 responses.Higher valueInfluenced by first- order performance
Meta-d’/d’ or Meta-d’-d’Rounis et al. (2010),
Maniscalco and Lau (2012)
Type-1 sensitivity expected for metacognitively ideal individual, calculated from observed secondorder responses, compared to observed type-1 sensitivity.Higher valueRequires extensive data or hierarchical Bayesian models for reliable interpretation

Measures used to represent metacognitive sensitivity from accuracy and confidence reports.

The limitations of these measures have been discussed by Fleming (2017) and Hoven et al. (2019).

Given that second-order cognition may be dissociated from first-order cognition (Fleming et al., 2010; De Gardelle and Mamassian, 2015; Desender et al., 2016), this suggests that metacognition does not depend only on the information from internal responses involved in first-order processes, but also on further cues such as the subjective experience of fluency or speed of the preceding decision (Fleming and Dolan, 2012). Since different processes are considered to contribute to first-order and metacognitive responses, it is possible that in some cases of mental disorder the cognitive processes for first-order responses are impaired while those for metacognitive ones are spared (Powers et al., 2017; Faivre et al., 2021), or vice versa (Chan et al., 2015; Berna et al., 2019). This dissociation also provides that an individual’s metacognitive sensitivity may be based on similar second-order processing across tasks from different first-order cognitive domains such as memory and perception. Some research findings have promoted a domain-specific conceptualization of metacognition insofar as it can be enhanced or impaired in one domain but not another (Baird et al., 2014; Fleming et al., 2014), which is supported by evidence of domain-specific patterns of activity in the prefrontal cortex as predicting metacognition, alongside a wider network of domain-general signals (Morales et al., 2018). However, reviews have reported inconclusive findings on the domain-specificity of metacognition (Rouault et al., 2018a; Vaccaro and Fleming, 2018).

Understanding the contribution of metacognition to symptoms of mental disorder depends on evaluating whether metacognitive sensitivity is specific to cognitive domain and whether variation exists at the level of local metacognitions, as opposed to their synthesis into global metacognitions (Seow et al., 2021). These insights will help to clarify at which level of metacognition treatments for psychopathology may function (Moritz and Lysaker, 2018). This systematic review and meta-analysis aims to evaluate whether local metacognitive monitoring varies in those with symptoms of mental disorder compared to those without such symptoms, and if so whether this relationship is found across cognitive domains.

2. Materials and methods

This systematic review and its meta-analyses have been conducted in line with PRISMA recommendations (Moher et al., 2009), in order to promote the replicability of findings and to increase confidence that conclusions regarding the research question are not based on a biased sample of available evidence. A protocol for the review was not registered but all other PRISMA procedural recommendations were followed as closely as possible and are described below.

2.1. Eligibility criteria

The eligibility of a study for inclusion in the review was determined according to the following pre-defined criteria. To be included, studies had to provide data on participants’ response accuracy in a behavioral task and its relationship to explicit reports of response confidence on a rating scale for after each trial. To reduce the possibility of bias in estimates of metacognition as discussed by Fleming and Lau (2014), studies were only included if performance on the behavioral task was equated or was shown to be not significantly different between groups. In addition, studies needed to have compared metacognition in participants grouped according to the presence and absence of psychiatric symptoms, i.e., in a binary fashion, rather than a continuous correlation between metacognitive sensitivity and symptom severity in a healthy population.

Studies that measured mental disorder or atypicality predominantly attributed to neurodevelopmental processes, neurodegenerative processes, neurological symptoms or brain injury and which can be dissociated from subjective mental well-being were excluded. Studies not published as English-language articles in peer-reviewed journals were excluded from this review.

2.2. Study search and selection strategy

Records of studies were sourced through searches of the online databases APA PsychInfo, PubMed, Scopus and Web of Science during October 2022. Further potentially relevant studies were then sourced by manually searching relevant review publications returned by the database searches, as well as citations from database-retrieved studies which had been identified as meeting the pre-defined inclusion criteria after full-text screening.

The choice of search terms was partly guided by those used by Hoven et al. (2019), who examined confidence in relation to psychopathology, with the addition of terms excluding studies based on questionnaire measures of attributive metacognition. The databases were searched for articles whose titles included any of the terms: “metacogniti*” or “metamemory” and either “psychiatr*” or “impulsiv*” or “compulsiv*” or “symptom*” or “depressi*” or “schizo*” or “OCD” or “addict*” or “substance*” or “eating” or “MDD” or “gambl*” or “anxi*” or “psychos*” or “disorder*,” but without “treatment*”, “training,” “therap*” or “belief*” in the abstract or title.

Duplicates were removed from the collection of article records produced by combining the four database searches and then the titles and abstracts of the records were screened in order to assess whether they could meet the pre-established inclusion criteria. Where it seemed possible that a study met inclusion criteria based on the title and abstract, the full text was retrieved and screened to confirm eligibility for inclusion. The reference sections of articles confirmed as eligible for inclusion in the review were searched for further potentially relevant articles. Where data necessary for effect size calculation was not available in the published article, the corresponding author named in the article was contacted with a request for the information required to calculate the effect size. For studies where no data was made available (or would have required substantial re-analyses) to the authors of this review, effect sizes published in other review articles (where available and compatible) were used so that they could be included in the meta-analysis, as referenced in the results section.

2.3. Data extraction and analysis

Information regarding study design variables and sample characteristics was extracted for each of the articles in the final selection. The relevant elements of this information were incorporated into a summary of methodological quality and risk of bias across studies. Separate meta-analyses were conducted for studies grouped according to whether symptoms of mental disorder present in their samples were related to psychosis, in order that heterogeneity between studies within each meta-analysis was kept to a level that permitted meaningful comparison. It was assumed that samples from different studies were independent where no evidence existed to suggest otherwise. For studies that measured metacognition across multiple groups with different symptom levels, the effect size calculated was based on the difference between the group with the lowest symptoms of psychopathology (usually the “healthy control” group) and the group with the highest. Where metacognition was measured in the same sample across different cognitive domains, data from the perceptual domain was entered into the meta-analysis, and if more than on perceptual domain was investigated, the visual perception data was included, since first order performance is easier to control in the perceptual domain and visual tasks have been the common method of experimental investigation of metacognition. Analysis of pre-calculated effect sizes was conducted in the IBM SPSS Statistics (Version 28.0, IBM Corp, 2021) using the continuous meta-analysis procedure with DerSimonian and Laird (1986) random-effects estimation including Hartung and Knapp (2001) adjustment to reflect uncertainty in between-study heterogeneity estimation as recommended by Deeks et al. (2019). This analysis produced statistics for heterogeneity of effect sizes, a summary effect size, moderator effects and publication bias for both groups of studies.

2.3.1. Individual effect sizes

Individual effect sizes were estimated for each study by calculating Hedges’ g, an effect size of standardized mean difference, produced by dividing the difference in group means by a value for standard deviation pooled across study groups, which is adjusted for bias arising from small sample size (Hedges and Olkin, 1985; Wilson, 2021). The studies’ effect sizes, with corresponding standard error (SE) were computed using the web-based effect-size calculator developed by Wilson (2001).

2.3.2. Heterogeneity

The presence of significant heterogeneity across study effect sizes was investigated by calculating Cochran’s Q statistic as well as the I2 statistic to reflect the percentage of the total variation due to between-study variability, as recommended by Higgins and Thompson (2002).

2.3.3. Summary effect size

To produce the summary effect size for each meta-analysis, the DerSimonian and Laird (1986) random-effects model was applied, weighting individual effect sizes by a combination of within and between-study variance to produce a summary effect size.

2.3.4. Moderator analysis

Given previous findings that metacognition is domain-specific (e.g., Fleming et al., 2014), a Q statistic was calculated as a test for homogeneity of studies grouped by cognitive domain of first-order task, to assess whether the cognitive task used in studies was linked to significant variability in effect sizes.

2.3.5. Publication bias

Risk of publication bias was assessed by producing a funnel plot of study effect sizes relative to their standard error, accompanied by calculation of Egger’s regression test (Egger et al., 1997) of funnel plot asymmetry.

3. Results

3.1. Systematic review of literature search results

The results of the study identification, screening and selection process are outlined in the PRIMSA diagram in Figure 2. At the end of this process, 20 studies from the literature search were determined eligible for inclusion in the review.

FIGURE 2

FIGURE 2

PRISMA flowchart representing the process producing the final selection of studies included in the review.

3.2. Study characteristics

The symptoms of mental disorder measured across the 20 studies from the literature search which met the eligibility criteria were more frequently linked to psychosis (n = 12), while studies measuring non-psychotic mental disorder (n = 8) included samples with symptoms of compulsivity (n = 4), addictive disorder (n = 3) and functional cognitive disorder (n = 1). Across studies of various symptom profiles, the majority used samples whose symptoms were present at a level warranting clinical diagnosis (n = 17). The three non-clinical studies used self-report questionnaire scores to group participants according to relative levels of psychopathology symptoms. Studies that used perceptual first-order tasks to assess metacognition (n = 12) equated first-order performance between groups and those that used non-perceptual tasks (n = 8) found performance to be non-significantly different without manipulation. Perceptual tasks were most frequently in the visual modality (n = 11), with only one study using an auditory task (Powers et al., 2017). Of the non-perceptual tasks employed, most assessed memory (n = 5), one involved general knowledge, one emotion discrimination and another response inhibition. Study design variables and sample sizes for the studies included in the review are presented in Table 4.

TABLE 4

ReferencesPsychop. participantsControl participantsMeasure of symptomsFirst-order taskPerformance comparisonMeasure of metacognition
Ben Shachar et al. (2013)2522Obsessive- compulsive tendencies measured by OCI-RGeneral knowledge: 2 alternative forced-choice recognitionNo differenceGoodman– Kruskall gamma coefficient
Berna et al. (2019)109DSM 4 schizophrenia diagnosisMemory: Autobiographical recognitionNo differenceMeta-d’-d’
Bhome et al. (2022)1442FCD diagnosis through specialist neuropsychiatry servicePerceptual: Visual discriminationEquatedMeta-d’/d’*
Chan et al. (2015)2220Schizotypal traits as measured by SPQ-BRResponse inhibition: Flanker taskNo differenceLow confident incorrect responses
Davies et al. (2018)3118DSM 4 FEP diagnosisPerceptual: Visual discriminationEquatedMeta-d’/d’
Eisenacher et al. (2015)2138FEP according to DSM 4 criteriaMemory: Word recognitionNo differenceConfidence gap**
Faivre et al. (2021)2120DSM 5 schizophrenia spectrum diagnosisPerceptual: Visual discriminationEquatedMeta-d’/d’*
Gaweda et al. (2018)2533DSM 4 FEP diagnosisMemory: Recognition of prior actionsNo differenceKnowledge Corruption Index**
Hauser et al. (2017)2020Trait compulsivity as measured by PI-WSURPerceptual: Visual discrimination Perceptual:EquatedMeta-d’/d’*
Hoven et al. (2022)2755Clinical assessment for gambling disorderVisual discriminationEquatedMeta-d’/d’*
Hoven et al. (2022)2855Clinical assessment for OCDPerceptual: Visual discriminationEquatedMeta-d’/d’*
Jia et al. (2020)3838Clinical assessment through positive and negative symptoms scores (PANSS)Perceptual: Visual discriminationEquatedAUROC2
Kircher et al. (2007)2719DSM 4 FEP diagnosisMemory: Word recognitionNo differenceRelationship between confidence and error responses
Koizumi et al. (2020)1718DSM 4 schizophrenia diagnosis.Perceptual: Visual detectionEquatedMeta-d’/d’
Moeller et al. (2016)813DSM 4 SUD diagnosisPerceptual: Visual discriminationEquatedMeta-d’/d’
Moritz et al. (2012)2329DSM 4 schizophrenia diagnosisEmotion discriminationNo differenceHigh confident incorrect responses
Powers et al. (2017)1515DSM 4 schizophrenia diagnosisPerceptual: Auditory detectionEquatedMeta-d’/d’*
Sadeghi et al. (2017)2324DSM 4 SUDPerceptual: VisualEquatedMeta-d’/d’
Tekcan et al. (2007)2527DSM 4 OCD diagnosisMemory: Semantic recognitionNo differenceGoodman– Kruskall gamma coefficient
Wright et al. (2020)4868ICD 10 FEP diagnosisPerceptual: VisualEquatedMeta-d’/d’

Characteristics of studies included in the review: ‘Psuchop. Participants’ refers to psychopathological participants.

First-order performance denoted as “equated” indicates that accuracy was manipulated to be equal between participants, whereas “no difference” indicates that accuracy was not significantly different between participants in the absence of specific experimental manipulation to produce this result. OCI-R, obsessive compulsive inventory - Revised; DSM, Diagnostic and Statistical Manual of Mental Disorders; SPQ-BR, Schizotypal Personality Questionnaire-Brief Revised; FEP, first episode psychosis; PI-WSUR, Padua Inventory-Washington State University Revision; PANSS, The Positive and Negative Syndrome Scale; OCD, obsessive-compulsive disorder; SUD, substance use disorder; ICD, International Statistical Classification of Diseases and Related Health Problems. Descriptions of the metacognitive measures can be found in Table 1. *Studies used hierarchical Bayesian modeling to generate group-level parameter estimates **Studies whose effect sizes are the values cited in Rouy et al. (2021).

3.3. Assessment of methodological quality and risk of bias

It has been demonstrated that measures of metacognitive sensitivity are compromised in validity if they can be influenced either by general confidence bias or first-order performance (Galvin et al., 2003). The unwanted influence of first-order performance has in some cases been addressed by equating this across participants, although this has been argued to inflate estimates of metacognitive ability (Rahnev and Fleming, 2019). The twelve studies included in the review employing tasks in the perceptual domain used performance-equating techniques, usually staircase-adjustment procedures, in order to make first-order performance comparable between participants. For the studies using non-perceptual tasks, which neither implemented such a procedure nor used model-based metrics that take into account type-1 d’, the assumption that metacognitive sensitivity measures are not confounded by first-order accuracy relied on follow-up analysis demonstrating nonsignificant performance difference between groups. In these instances, the degree of performance accuracy potentially varies more between the groups of participants and so may produce systematic differences in estimates of metacognitive sensitivity compared to studies equating performance. Where authors did not report results using the meta-d’/d’ or meta-d’-d’ metrics, which was the case for eight of the included studies, insufficient data was available to enable the computation of these measures of sensitivity relative to first-order performance in studies.

3.4. Metacognitive sensitivity in the presence of psychosis-related symptoms

3.4.1. Individual and summary effect sizes

A meta-analysis was conducted to compare effect sizes for differences in metacognitive sensitivity between samples with and without psychosis-related symptoms, using the Hedges’ g form of standardized mean difference as an estimate of effect size. The individual effect sizes, their 95% Confidence Interval (CI) and the weight with which they contribute to the summary effect size are represented in the forest plot in Figure 3. The summary effect size can be considered to be of medium magnitude based on guidelines recommended for a related measure of standardized mean difference, Cohen’s d (Cohen, 1988), and indicated lower metacognitive sensitivity in those with psychosis-related symptoms compared to those not displaying psychopathology (g = −0.53, 95% CI = −0.71, −0.35). A t-test indicated that the observed summary effect was very unlikely to have arisen by chance (t(11) = −6.34, p < 0.001).

FIGURE 3

FIGURE 3

Forest plot of the distribution of Hedges’ g effect sizes for metacognitive sensitivity across studies of samples with psychosis-related symptoms, based on a random- effects analysis, displaying effects by arranged sub-group of task domain, which was either perceptual (1) or non-perceptual (2). Lower metacognitive sensitivity in those with psychosis-related symptoms is indicated by a negative effect size. The summary effect size is indicated by a diamond marker, underneath the individual study effect sizes.

3.4.2. Heterogeneity

The existence of heterogeneity between study effect sizes was tested by calculating the Q statistic, which failed to indicate significant heterogeneity (Q = 11.16, df = 11, p = 0.43). The proportion of variability between effect sizes not attributable to sampling error was quantified using the I2 statistic (Higgins and Thompson, 2002) and this was found to be negligible ( < 1%).

3.4.3. Moderator analysis

Although there was no evidence for heterogeneity across studies included in the analysis, a planned investigation was conducted as to whether variation in effect sizes was related to the cognitive domain of the first-order task used, given previous findings that metacognition is domain-specific (e.g., Fleming et al., 2014). The Q statistic calculated as a test for homogeneity of studies grouped by cognitive domain of first-order task did not provide evidence of a moderating effect of cognitive domain task on measures of metacognitive sensitivity (Q = 2.32, df = 1, p = 0.13).

3.4.4. Publication bias

Publication bias was examined by assessing the symmetry of the distribution of included effect sizes in terms of their precision. Figure 4 shows a funnel plot constructed around the summary effect size and represents the area within which 95% of studies should fall in the absence of heterogeneity and biases (Deeks et al., 2019). The lack of marked asymmetry in the funnel plot as depicted in Figure 4 suggested there is no significant publication bias for results in this area, which was supported by a non-significant value for Egger’s regression test (Egger et al., 1997) of funnel plot asymmetry (t(11) = −0.77, p = 0.94).

FIGURE 4

FIGURE 4

Funnel plot of the distribution of effect sizes by their standard error for studies of metacognition in the presence of psychosis-related symptoms. The vertical line indicates the value of the summary effect size. The area of the graph within the triangle represents the values which samples have 95% probability of showing if variance is homogeneous. Funnel plot of the distribution of effect sizes by their standard error for studies of metacognition in the presence of psychosis-related symptoms.

3.5. Metacognitive sensitivity in the presence of non-psychotic symptoms of mental disorder

3.5.1. Individual and summary effect sizes

A second meta-analysis was conducted for effect sizes across studies comparing metacognitive sensitivity in samples with non-psychosis-related symptoms of mental disorder and healthy individuals. The resulting summary effect size, reflecting the weighted averages of individual studies’ effect sizes, is displayed in Figure 5. Unlike the meta-analysis for studies investigating metacognition in individuals with psychosis-related symptoms, the summary effect size for this group of studies did not provide evidence of an overall difference in metacognition compared to those without symptoms of mental disorder (g = −0.24, 95% CI = −0.56, 0.08). A t-test indicated that the observed summary effect was not significant t(7) = −1.78, p = 0.12).

FIGURE 5

FIGURE 5

Forest plot of the distribution of effect sizes for metacognitive sensitivity across studies of samples with non-psychotic symptoms of mental disorder, based on a random-effects analysis, displaying effects by arranged sub-group of task domain, which was either perceptual (1) or non-perceptual (2). Lower metacognitive sensitivity in those with non-psychotic symptoms of mental disorder is indicated by a negative effect size. The summary effect size is indicated by a diamond marker, underneath the individual study effect sizes.

3.5.2. Heterogeneity

The existence of heterogeneity between study effect sizes was tested by calculating the Q statistic, which failed to indicate significant heterogeneity (Q = 12.43, df = 7 p = 0.09). The proportion of variability between effect sizes not attributable to sampling error was quantified using the I2 statistic (Higgins and Thompson, 2002); this did indicate that a moderate degree of heterogeneity existed between studies (I2 = 44%; Deeks et al., 2019).

3.5.3. Publication bias

Publication bias was by examined by assessing the symmetry of the distribution of included effect sizes in terms of their precision. Figure 6 shows a funnel plot constructed around the summary effect size and represents the area within which 95% of studies should fall in the absence of heterogeneity and biases (Deeks et al., 2019). The asymmetry of distribution in the funnel plot as depicted in Figure 6 suggested there may be a publication bias for results in this area, although a non-significant value was found for Egger’s regression test (Egger et al., 1997) of funnel plot asymmetry (t(7) = −1.88, p = 0.11). It has, however, been argued that both funnel plot symmetry may fail to accurately reflect publication bias where the number of studies included is low (Begg and Mazumdar, 1994), as it the case for this meta-analysis.

FIGURE 6

FIGURE 6

Funnel plot of effect sizes by standard error for studies of metacognition in the presence of non-psychotic symptoms of mental disorder. The vertical line indicates the value of the summary effect size. The area of the graph within the triangle represents the values which samples have 95% probability of showing if variance is homogeneous.

3.5.4. Moderator analysis

As for studies of psychosis-related symptoms, a planned investigation was conducted as to whether effect sizes varied for studies measuring metacognition using tasks from perceptual or non-perceptual cognitive domains. Although there was no evidence for heterogeneity across studies included in the analysis, a planned investigation was conducted as to whether variation in effect sizes was related to the cognitive domain of the first-order task used, given previous findings that metacognition is domain-specific (eg., Fleming et al., 2014). As with the previous meta-analysis, the Q statistic calculated as a test for homogeneity of studies grouped by cognitive domain of first-order task did not provide evidence of a moderating effect of cognitive domain task on measures of metacognitive sensitivity (Q = 0.52, df = 1, p = 0.47).

4. Discussion

4.1. Summary

This systematic review and meta-analysis aimed to establish whether metacognitive sensitivity differs between those with and without symptoms of mental disorder. Metacognitive sensitivity was defined as the ability to discriminate first-order response accuracy through reports of confidence with respect to individual responses in a task. We further sought to test whether this depended on the domain of the first-order cognitive task used to measure metacognitive sensitivity, in particular whether there were any differences between studies employing perceptual versus non-perceptual tasks. The results showed that metacognitive sensitivity was significantly reduced in those with symptoms of mental disorder related to psychosis but only a (nonsignificant) trend could be observed in those with symptoms related to obsessive-compulsive disorder (OCD), substance use disorder (SUD) or functional cognitive disorder (FCD). The effects found are based on studies using tasks requiring first-order cognition in different domains, here categorized as perceptual versus non-perceptual. No evidence was found that the cognitive domain moderated metacognitive sensitivity, since the between-group Q statistic for studies grouped as perceptual and non-perceptual did not reach significance in either meta-analysis.

4.2. Relation of findings to existing literature

The results of the meta-analyses in this review converge to some extent with the conclusions drawn by Hoven et al. (2019) from a qualitative review of research into metacognitive bias and sensitivity in populations with diverse symptoms of mental disorder, also conducted across a variety of first-order cognitive domains. Hoven et al. found that the balance of evidence suggested impairment of metacognitive sensitivity in those with subclinical obsessive-compulsive tendencies as well as in those with psychosis-related symptoms. However, research relating to symptoms of clinical OCD, addiction, anxiety and depression was either absent or provided mixed evidence for impairment of metacognitive sensitivity. The difficulty in drawing conclusions from many of the studies about metacognitive sensitivity in the review by Hoven et al. is due to the possibility that they were confounded by confidence biases and first order performance. The studies in the current review have equated or not significantly different first-order performance, reducing the possibility of bias in estimating metacognitive sensitivity, although studies employing non-perceptual tasks have measures with higher risk of being influenced by first-order performance.

The research of Rouault et al. (2018b) specifically related variability in metacognitive efficiency, a measure of metacognitive sensitivity relative to performance, across symptoms ascribed to different diagnostic categories, which has parallels with the findings of the current review. Rouault, Seow et al. found that metacognition was predicted in a general population sample by transdiagnostic symptom dimensions derived through factor analysis of individual items from a range of psychiatric questionnaires, replicating a latent structure across diagnostic categories originally obtained by Gillan et al. (2016). Metacognitive efficiency was found to be negatively correlated with scores in one kind of symptom dimension, characterized by compulsivity and intrusive thoughts, while it was positively correlated with scores on another dimension, characterized by anxiety and depression. While the indication of a negative association between metacognitive efficiency and the compulsivity-intrusive thought dimension is in agreement with the direction of the trend found in the current meta-analysis of studies involving compulsivity and addiction, although the overall effect did not reach significance. However, no research involving symptoms of anxiety or depression met criteria for inclusion in the current review. Based on the research by Rouault, Seow, et al. and other findings that fail to demonstrate any impairment in sensitivity in relation to symptoms of depression (Herskovik et al., 1986; Fieker et al., 2016; Moses-Payne et al., 2019), it seems unlikely that the effect found for those symptoms of psychopathology included in this analysis can be generalized to symptoms of anxiety or depression.

The results in the current review follow the direction of findings in another recent meta-analysis by Rouy et al. (2021) synthesizing research in samples with diagnoses of schizophrenia spectrum disorders, which found a strong overall effect size (g = −0.57) for reduced metacognitive sensitivity in those with schizophrenia compared to control participants. This effect size was however based on a selection of studies that did not universally control for first-order performance. Rouy et al. performed further analyses which indicated that the magnitude of the summary effect size was substantially reduced when estimated only from those studies actively matching first-order performance across participants. The reduced effect size found by Rouy et al. (2021) is at odds with the larger effect found for symptoms of psychosis in the current review, which also included only studies that equated first-order performance, or which demonstrated no significant difference between groups’ performance. The larger effect size for studies of psychosis-related symptoms in the current analysis may be attributable to the inclusion of those studies in which performance was not actively equated, only not significantly different.

4.3. Associations of metacognition with neuroanatomical and functional activation differences

Attempts to explain interindividual differences in metacognition have assessed metacognitive sensitivity in relation to neuroanatomical features and task-related activation. Fleming et al. (2010) found that strategic metacognition was predicted by individual differences in the structure of anterior prefrontal cortex, while the dorsolateral prefrontal cortex has been causally implicated in metacognition accuracy through application of repetitive transcranial magnetic stimulation (Rounis et al., 2010). Since structural and functional atypicality in areas of the prefrontal cortex have been linked to a range of mental disorders (Koenigs and Grafman, 2009; Goldstein and Volkow, 2011), the evidence for the role of prefrontal areas in metacognitive processing contributes to an expectation of impaired metacognitive sensitivity across different kinds of mental disorder. This neuroanatomical link is supported by studies in the current review; Moeller et al. (2016) found that decreased gray matter volume in the anterior prefrontal cortex predicted the degree of perceptual metacognitive impairment in substance use disorder and research by Jia et al. (2020) indicated that differences in schizophrenia patients’ metacognition were linked to frontoparietal hypoactivity. However, the absence of an association between impaired perceptual metacognitive sensitivity in psychosis and gray matter volume reduction in the prefrontal cortex in the study by Davies et al. (2018) means that further research is required in order to clarify the contribution of this region to metacognitive sensitivity. Furthermore, a review of neuroanatomical associations identified in relation to metacognition has suggested that involvement of substrates in metacognition may depend on the cognitive domain in which it is exercised, and that metamemory performance may be more reliant on structures outside of the prefrontal cortex (Vaccaro and Fleming, 2018).

4.4. Domain-generality of metacognitive performance

A recent review has investigated whether individuals’ metacognitive performance is correlated across distinct cognitive domains and found that evidence was inconclusive regarding the domain-general nature of metacognition (Rouault et al., 2018a). In the current review, those studies using tasks in the perceptual domain were able to actively equate performance between groups, while studies using non-perceptual tasks were not. This has the implication that any differences in effects between sub-groups distinguished by task domain may reflect not only the influence of domain but also that of equating first-order performance. The lack of significant variability in effect sizes between the sub-group of studies using perceptual tasks and equating performance and the subgroup using non-perceptual tasks without equating performance can be interpreted in two possible ways. Firstly, it is possible that neither task domain nor control of performance significantly influence effect size. While this does not refute that metacognitive capacity may differ across first-order cognitive domains, it would imply that its impairment is observed to the same extent across domains in those with psychiatric symptoms. Secondly, it is possible that different degrees of metacognitive impairment do exist for different first-order domains but that this is offset by the influence of the co-varying performance manipulation. It should also be considered that differences in metacognitive sensitivity between cognitive domains may be obscured by the fact that there is also variation in terms of the stimuli and measures used within domains. Stimulus-level variables have been shown to produce different estimates of metacognition, as in the case of spatial frequency (Koizumi et al., 2020) for perceptual metacognition or episodic rather than semantic content (Tekcan et al., 2007; Berna et al., 2019) for metamemory, while different metrics within domains also reduce the clarity of cross-domain comparisons. The heterogeneous combinations of design and measurement features across studies prevent any clear conclusions regarding the influence of any individual factor on estimates of metacognitive sensitivity.

4.5. Interpreting measures of metacognitive performance

Some challenges have been raised with respect to conclusions drawn from the measures of metacognitive sensitivity employed in the studies reviewed, which have implications for the interpretation of the group differences observed. Metacognitive monitoring has been conceived as involving the processing of internal evidence arising from first-order cognition (Fleming et al., 2012) and that equating first-order performance provides that the evidence available for metacognitive processing is also equated, enabling a measure of true metacognitive sensitivity. It has been argued by Paulewicz et al. (2020) that neither equating first-order performance nor implementing model-based measures of second-order sensitivity that are formulated relative to first-order sensitivity are sufficient to isolate metacognitive performance from first-order processes. Paulewicz et al. (2020) raise this concern in the context of a causal analysis framework, taking into account that metacognition may involve the monitoring or regulation of multiple interacting stages which constitute first-order processing for a given task (Flavell, 1979). The authors argue that equalizing first-order accuracy may not standardize all first-order processes or metacognitive regulatory activity, which means that intended measures of metacognitive sensitivity may still reflect differences in first-order processes or metacognitive regulation where first-order performance is the same. Following this reasoning, the effect for differences in metacognitive measures between groups may require a more limited interpretation as a difference in the statistical relation between accuracy and confidence, rather than differential operation of a metacognitive monitoring process.

4.6. Correspondence of local metacognition to global metacognition and mental health

Although it is difficult to specifically isolate metacognitive monitoring through empirical measures, it is useful to compare the relation of second-order judgments to standardized behavioral measures of cognition. This operationalization of metacognition, based on relatively discrete cognitive functions involved in task performance, allows more direct interindividual comparison of second-order judgments than attributive measures of metacognition, which are based on synthesizing diverse sets of cognitive processes across contexts (Lysaker et al., 2018). Nevertheless, the complexity of cognitive processes varies even across task-based metacognition, given that some studies involve tasks which require only low-level stimulus processing in terms of contrast or global motion (Fleming et al., 2010; Hauser et al., 2017) whereas others involve of stimuli like facial expressions which are more likely to invoke multiple or high-order cognitive processes (Moritz et al., 2012). These latter examples may assess cognitive processes more comparable to those represented by measures of global metacognition such as the Metacognitions Questionnaire (MCQ; Cartwright-Hatton and Wells, 1997). The correspondence between local and global forms of metacognition and their contribution to mental health is receiving increased attention in recent research (for a review see Seow et al., 2021). Although Lee et al. (2021) found that global metacognition is based on integrating instances of local metacognition, the information was not optimally integrated and influenced by additional variables. Chan et al. (2015) compared both local and global metacognition in individuals with high and low schizotypy. Chan et al. (2015) found that groups differed significantly only in measures of global metacognition as reported in the MCQ, not local metacognition. More recently, Bhome et al. (2022) also found evidence of a dissociation between local and global metacognition, reporting intact local metacognition in FCD patients but reduced global metacognition. Further comparison of local and global metacognition would help to clarify the extent to which these are differentially impacted in psychopathology.

To determine the functional relevance of task-based measures of metacognitive ability to the assessment or treatment of those demonstrating symptoms of psychopathology it is important to consider whether these measures specifically predict symptom severity or broader functioning. Where significant overall reductions in metacognitive sensitivity have been found for samples with psychiatric symptoms relative to control groups, metacognitive performance has predicted symptom severity in some cases (Eisenacher et al., 2015; Jia et al., 2020) but not others (Kircher et al., 2007; Davies et al., 2018; Gaweda et al., 2018). As the current review included only three studies for sub-clinical level psychiatric symptoms, it was not viable to use sub-group analysis to evaluate whether the extent of metacognitive impairment depended on symptom severity reaching a clinical threshold. Where metacognition is assessed in relation to symptoms that present at a prodromal level or within an initial episode of psychosis, the extent of metacognitive impairment may guide clinical intervention if this is able to predict likelihood of future episodes or recovery prognosis (Hauser et al., 2017). At present, no study fulfilling our criteria has implemented a longitudinal design to establish the potential of metacognitive assessment in this respect.

4.7. Limitations of the current review

The current review’s conclusions regarding metacognitive impairment in psychopathology may be considered limited in the sense that they only reflect research involving between-group comparisons, so overlook findings relating continuous measures of psychiatric symptoms to metacognitive sensitivity (Rouault et al., 2018b; Moses-Payne et al., 2019). Conclusions of impaired metacognitive sensitivity are based only on groups with symptoms of psychosis, compulsivity, addiction or FCD as only these met pre-defined eligibility criteria regarding experimental design. For this reason, the results of analyses in this review are not informative with respect to other kinds of psychiatric symptoms. A factor that limits the conclusions that can be drawn regarding effects for the specific disorders in the reviewed studies is that only measures of the primary symptom(s) of interest characterizing the groups are reported, with a few exceptions (Eisenacher et al., 2015; Faivre et al., 2021). Given the presence of co-morbidity among many psychiatric diagnoses, and growing evidence for the existence transdiagnostic symptom clusters (Rouault et al., 2018b), it is appropriate to assess the potential influence on outcome measures of other symptoms beyond those primarily characterizing a sample. This is particularly relevant for symptoms of depression, which may vary considerably between participants with a similar psychiatric diagnosis as well as between participants assigned to an asymptomatic or low-symptom control group (Rouy et al., 2021). Where studies do not determine the presence of secondary symptoms or assess the relation of these to the outcome measures, their contribution to metacognitive differences attributed to the primary disorder cannot be evaluated.

Potential sources of bias in the studies included for review relate to the use of outcome measures that are vulnerable to first-order performance and confidence bias confounds (Fleming and Lau, 2014). A risk of confidence bias impacting measures of metacognitive sensitivity is found in those studies in non-perceptual domains comparing group sensitivity with measures based on the absolute number of incorrect responses given with high or low confidence or confidence-accuracy correlation (Masson and Rotello, 2009), a consideration particularly relevant to the current review insofar as this has been found to be heightened in some psychiatric populations (Hoven et al., 2019). A clearer understanding of the relationship between psychiatric symptoms and metacognitive sensitivity could be reached if future studies increase use of measures such meta-d’/ d’ which are not influenced by response bias and standardized in terms of first-order performance, allowing findings to be more easily related across samples and task procedures reliant on cognition in different domains (Fleming, 2017). Finally, it is important to note that since the studies included for review were all cross-sectional in nature, it is not possible to arrive at conclusions regarding the causal role of metacognitive impairments in producing associated symptoms.

5. Conclusion

This systematic review and meta-analysis of findings from research into metacognitive monitoring in individuals with psychiatric symptoms has provided evidence that metacognitive sensitivity is reduced in populations with symptoms related to psychosis. The overall effect found for dysfunction of metacognitive monitoring in those demonstrating features of psychotic psychopathology suggests that this, alongside particular impairments in first-order processes depending on the symptom profile (Hauser et al., 2017; Gaweda et al., 2018), may be implicated in the manifestation of these symptoms. There was however no conclusive evidence of reduced metacognitive sensitivity in those with non-psychotic symptomatology. No evidence was found to suggest second-order performance was dependent on specific to first-order cognitive domain in either meta-analysis. Future research using longitudinal designs may clarify the role of metacognitive sensitivity in the prognosis of psychiatric symptoms and mental disorder, which can guide decisions to recommend interventions of the kind which have been found to improve metacognitive sensitivity in healthy individuals (Baird et al., 2014) for those with psychiatric symptoms. A further avenue of research to clarify the functional relevance of metacognitive impairment could investigate how metacognitive sensitivity affects consequent cognition and behavior, which has as yet received limited attention (Ben Shachar et al., 2013) and which may also clarify how local and global forms of metacognition are related (Chan et al., 2015).

Statements

Data availability statement

The original contributions presented in this study are included in this article/supplementary material, further inquiries can be directed to the corresponding authors.

Author contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work, and approved it for publication.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1

    BairdB.MrazekM. D.PhillipsD. T.SchoolerJ. W. (2014). Domain-specific enhancement of metacognitive ability following meditation training.J. Exp. Psychol. General14319721979. 10.1037/a0036882

  • 2

    BeggC. B.MazumdarM. (1994). Operating characteristics of a rank correlation test for publication bias.Biometrics5010881101. 10.2307/2533446

  • 3

    Ben ShacharA.LazarovA.GoldsmithM.MoranR.DarR. (2013). Exploring metacognitive components of confidence and control in individuals with obsessive-compulsive tendencies.J. Behav. Ther. Exp. Psychiatry44255261. 10.1016/j.jbtep.2012.11.007

  • 4

    BernaF.ZouF.DanionJ. M.KwokS. C. (2019). Overconfidence in false autobiographical memories in patients with schizophrenia.Psychiatr. Res.279374375. 10.1016/j.psychres.2018.12.063

  • 5

    BhomeR.McWilliamsA.PriceG.PooleN. A.HowardR. J.FlemingS. M.et al (2022). Metacognition in functional cognitive disorder.Brain Commun.4:fcac041. 10.1093/braincomms/fcac041

  • 6

    BlikstedV.SamuelsenE.SandbergK.BibbyB. M.OvergaardM. S. (2017). Discriminating between first- and second-order cognition in first-episode paranoid schizophrenia*.Cogn. Neuropsychiatr.2295107. 10.1080/13546805.2016.1268954

  • 7

    Cartwright-HattonS.WellsA. (1997). Beliefs about worry and intrusions: the Meta-Cognitions Questionnaire and its correlates.J. Anxiety Disord.11279296. 10.1016/S0887-6185(97)00011-X

  • 8

    ChanC. C.SpencerC. C.WestC.ViegasC.BedwellJ. S. (2015). Metacognitive processes in psychometrically defined schizotypy.Psychiatry Res.230279286. 10.1016/j.psychres.2015.09.006

  • 9

    ChanK. K. S. (2016). Associations of symptoms, neurocognition, and metacognition with insight in schizophrenia spectrum disorders.Comprehen. Psychiatry656369. 10.1016/j.comppsych.2015.09.009

  • 10

    ClarkeF. R.BirdsallT. G.TannerJ. (1959). Two types of ROC curves and definitions of parameters.J. Acoustical Soc.Am.31629630.

  • 11

    CohenJ. (1988). Statistical Power Analysis for the Behavioral Sciences.New York, NY: Routledge Academic.

  • 12

    CooperM. J.OsmanS. (2007). Metacognition in body dysmorphic disorder–A preliminary explorationJ. Cogn. Psychother.21148155. 10.1891/088983907780851568

  • 13

    DavidA. S.BedfordN.WiffenB.GilleenJ. (2012). Failures of metacognition and lack of insight in neuropsychiatric disorders Failures of metacognition and lack of insight in neuropsychiatric disorders.R. Soc. Open Sci.36713791390. 10.1098/rstb.2012.0002

  • 14

    DaviesG.GreenwoodK. (2020). A meta-analytic review of the relationship between neurocognition, metacognition and functional outcome in schizophrenia.J. Ment. Health29496505. 10.1080/09638237.2018.1521930

  • 15

    DaviesG.RaeC. L.GarfinkelS. N.SethA. K.MedfordN.CritchleyH. D.et al (2018). Impairment of perceptual metacognitive accuracy and reduced prefrontal grey matter volume in first-episode psychosis.Cogn. Neuropsychiatry23165179. 10.1080/13546805.2018.1444597

  • 16

    De GardelleV.MamassianP. (2015). Weighting mean and variability during confidence judgments.PLoS One10:e0120870. 10.1371/journal.pone.0120870

  • 17

    DeeksJ. J.HigginsJ. P. T.AltmanD. G. (2019). “Analysing data and undertaking meta-analyses,” in Cochrane Handbook for Systematic Reviews of Interventions, edsHigginsJ. P. T.ThomasJ.ChandlerJ.CumpstonM.LiT.PageM. J.et al (Hoboken: John Wiley & Sons).

  • 18

    DerSimonianR.LairdN. (1986). Meta-analysis in clinical trials.Contro. Clin. Trials7177188. 10.1016/0197-2456(86)90046-2

  • 19

    DesenderK.Van OpstalF.HughesG.Van den BusscheE. (2016). The temporal dynamics of metacognition: Dissociating task-related activity from later metacognitive processes.Neuropsychologia825464. 10.1016/j.neuropsychologia.2016.01.003

  • 20

    EggerM.Davey SmithG.SchneiderM.MinderC. (1997). Bias in meta-analysis detected by a simple, graphical test.BMJ315629634. 10.1136/bmj.315.7109.629

  • 21

    EisenacherS.RauschF.AinserF.MierD.VeckenstedtR.SchirmbeckF.et al (2015). Investigation of metamemory functioning in the at-risk mental state for psychosis.Psychol. Med.4533293340. 10.1017/S0033291715001373

  • 22

    EvansS.AzzopardiP. (2007). Evaluation of a ‘bias-free’ measure of awareness.Spatial Vis.206177. 10.1163/156856807779369742

  • 23

    FaivreN.RogerM.PereiraM.de GardelleV.VergnaudJ. C.PasserieuxC.et al (2021). Confidence in visual motion discrimination is preserved in individuals with schizophrenia.J. Psychiatry Neurosci.466573. 10.1503/jpn.200022

  • 24

    FiekerM.MoritzS.KotherU.JelinekL. (2016). Emotion recognition in depression: An investigation of performance and response confidence in adult female patients with depression.Psychiatry Res.242226232. 10.1016/j.psychres.2016.05.037

  • 25

    FlavellJ. H. (1979). Metacognition and cognitive monitoring: A new area of cognitive-developmental inquiry.Am. Psychol.34906911.

  • 26

    FlemingS. M. (2017). HMeta-d: hierarchical Bayesian estimation of metacognitive efficiency from confidence ratings.Neurosci. Conscious.2017114. 10.1093/nc/nix007

  • 27

    FlemingS. M.DolanR. J. (2012). The neural basis of metacognitive ability.Biol. Sci.36713381349. 10.1098/rstb.2011.0417

  • 28

    FlemingS. M.DolanR. J.FrithC. D. (2012). Metacognition: Computation, biology and function.Biol. Sci.36712801286. 10.1098/rstb.2012.0021

  • 29

    FlemingS. M.LauH. C. (2014). How to measure metacognition.Front. Hum. Neurosci.8:19. 10.3389/fnhum.2014.00443

  • 30

    FlemingS. M.RyuJ.GolfinosJ. G.BlackmonK. E. (2014). Domain-specific impairment in metacognitive accuracy following anterior prefrontal lesions.Brain13728112822. 10.1093/brain/awu221

  • 31

    FlemingS. M.WeilR. S.NagyZ.DolanR. J.ReesG. (2010). Relating introspective accuracy to individual differences in brain structure.Science32915411543. 10.1126/science.1191883

  • 32

    GalvinS. J.PoddJ. V.DrgaV.WhitmoreJ. (2003). Type 2 tasks in the theory of signal detectability: Discrimination between correct and incorrect decisions.Psychonomic Bull. Rev.10843876. 10.3758/BF03196546

  • 33

    GawedaŁ.LiE.LavoieS.WhitfordT. J.MoritzS.NelsonB. (2018). Impaired action self-monitoring and cognitive confidence among ultra-high risk for psychosis and first-episode psychosis patients.Eur. Psychiatr.476775. 10.1016/j.eurpsy.2017.09.003

  • 34

    GillanC. M.KosinskiM.WhelanR.PhelpsE. A.DawN. D. (2016). Characterizing a psychiatric symptom dimension related to deficits in goaldirected control.ELife5124. 10.7554/eLife.11305

  • 35

    GoldsteinR. Z.VolkowN. D. (2011). Dysfunction of the prefrontal cortex in addiction: Neuroimaging findings and clinical implications.Nat. Rev. Neurosci.12652669. 10.1038/nrn3119

  • 36

    GoodmanL. A.KruskalW. H. (1954). Measures of association for cross classifications.J. Am. Stat. Assoc.49732769. 10.2307/2281536

  • 37

    HartungJ.KnappG. (2001). A refined method for the meta-analysis of controlled clinical trials with binary outcome.Stat.Med.2038753889. 10.1002/sim.1009

  • 38

    HauserT. U.AllenM.ReesG.DolanR. J.BullmoreE. T.GoodyerI.et al (2017). Metacognitive impairments extend perceptual decision making weaknesses in compulsivity.Sci. Rep.7110. 10.1038/s41598-017-06116-z

  • 39

    HedgesL. V.OlkinI. (1985). Statistical methods for meta-analysis.New York: Academic Press.

  • 40

    HerskovikJ. E.KietzmanM. L.SuttonS. (1986). Visual flicker in depression: Response criteria, confidence ratings and response times.Psychol. Med.16187197. 10.1017/S0033291700002622

  • 41

    HeyesC.BangD.SheaN.FrithC. D.FlemingS. M. (2020). Knowing Ourselves Together: The Cultural Origins of Metacognition.Trends Cogn. Sci.24349362. 10.1016/j.tics.2020.02.007

  • 42

    HigginsJ. P. T.ThompsonS. G. (2002). Quantifying heterogeneity in a meta-analysis.Statist. Med.2115391558. 10.1002/sim.1186

  • 43

    HovenM.de BoerN. S.GoudriaanA. E.DenysD.LebretonM.van HolstR. J.et al (2022). Metacognition and the effect of incentive motivation in two compulsive disorders: Gambling disorder and obsessive-compulsive disorder.Psychiatry Clin. Neurosci.76437449. 10.1111/pcn.13434

  • 44

    HovenM.LebretonM.EngelmannJ. B.DenysD.LuigjesJ.van HolstR. J. (2019). Abnormalities of confidence in psychiatry: an overview and future perspectives.Trans. Psychiatry9118. 10.1038/s41398-019-0602-7

  • 45

    IBM Corp (2021). IBM SPSS Statistics for Windows, Version 28.0.Armonk, NY: IBM Corp.

  • 46

    JiaW.ZhuH.NiY.SuJ.XuR.JiaH.et al (2020). Disruptions of frontoparietal control network and default mode network linking the metacognitive deficits with clinical symptoms in schizophrenia.Hum. Brain Mapp.4114451458. 10.1002/hbm.24887

  • 47

    KircherT. T. J.KochK.StottmeisterF.DurstV. (2007). Metacognition and reflexivity in patients with schizophrenia.Psychopathology40254260. 10.1159/000101730

  • 48

    KoenigsM.GrafmanJ. (2009). The functional neuroanatomy of depression: Distinct roles for ventromedial and dorsolateral prefrontal cortex.Behav. Brain Res.201239243. 10.1016/j.bbr.2009.03.004

  • 49

    KoizumiA.HoriT.ManiscalcoB.HayaseM.MishimaR.KawashimaT.et al (2020). Atypical spatial frequency dependence of visual metacognition among schizophrenia patients.NeuroImage2718. 10.1016/j.nicl.2020.102296

  • 50

    LeeA. L. F.de GardelleV.MamassianP. (2021). Global visual confidence.Psychonomic Bull. Rev.2812331242. 10.3758/s13423-020-01869-7

  • 51

    LysakerP. H.GagenE.MoritzS.SchweitzerR. D. (2018). Metacognitive approaches to the treatment of psychosis: A comparison of four approaches.Psychol. Res. Behav. Manag.11341351. 10.2147/PRBM.S146446

  • 52

    ManiscalcoB.LauH. (2012). A signal detection theoretic approach for estimating metacognitive sensitivity from confidence ratings.Conscious. Cogn.21422430. 10.1016/j.concog.2011.09.021

  • 53

    MassonM. E. J.RotelloC. M. (2009). Sources of bias in the Goodman–Kruskal gamma coefficient measure of association: Implications for studies of metacognitive processes.J. Exp.35509527. 10.1037/a0014876

  • 54

    MoritzS.WoodwardT.RuffC. (2003). Source monitoring and memory confidence in schizophrenia.Psychol. Med.33131139. 10.1017/S0033291702006852

  • 55

    MoritzS.WoodwardT. S.CuttlerC.WhitmanJ. C.WatsonJ. M. (2004). False memories in schizophrenia.Neuropsychology18276283. 10.1037/0894-4105.18.2.276

  • 56

    MoellerS. J.FlemingS. M.GanG.ZilverstandA.MalakerP.d’Oleire UquillasF.et al (2016). Metacognitive impairment in active cocaine use disorder is associated with individual differences in brain structure.Eur. Neuropsychopharmacol.26653662. 10.1016/j.euroneuro.2016.02.009

  • 57

    MoherD.LiberatiA.TetzlaffJ.AltmanD. G.AltmanD.AntesG.et al (2009). Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement.PLoS Med.6:e1000097. 10.1371/journal.pmed.1000097

  • 58

    MoralesJ.LauH.FlemingS. M. (2018). Domain-general and domain-specific patterns of activity supporting metacognition in human prefrontal cortex.J. Neurosci.3835343546. 10.1523/JNEUROSCI.2360-17.2018

  • 59

    MoritzS.LysakerP. H. (2018). Metacognition - What did James H. Flavell really say and the implications for the conceptualization and design of metacognitive interventions.Schizophr. Res.2012026. 10.1016/j.schres.2018.06.001

  • 60

    MoritzS.WoznicaA.AndreouC.KotherU. (2012). sponse confidence for emotion perception in schizophrenia using a continuous facial sequence task.Psychiatry Res.200202207. 10.1016/j.psychres.2012.07.007

  • 61

    MorrisonA. P.FrenchP.WellsA. (2007). Metacognitive beliefs across the continuum of psychosis: Comparisons between patients with psychotic disorders, patients at ultra-high risk and non-patients.Behav. Res. Ther.4522412246. 10.1016/j.brat.2007.01.002

  • 62

    Moses-PayneM. E.RollwageM.FlemingS. M.RoiserJ. P. (2019). Postdecision Evidence Integration and Depressive Symptoms.Front. Psychiatry10:639. 10.3389/fpsyt.2019.00639

  • 63

    NeisserU. (1967). Cognitive psychology.New York: Appleton-Century-Crofts.

  • 64

    NelsonT. O.LeonesioR. J.LandwehrR. S.NarensL. (1986). A comparison of three predictors of an individual’s memory performance: The individual’s feeling of knowing versus the normative feeling of knowing versus base-rate item difficulty.J. Exp. Psychol.12279287. 10.1037/0278-7393.12.2.279

  • 65

    PaulewiczB.SiedleckaM.KoculakM. (2020). Confounding in Studies on metacognition: A preliminary causal analysis framework.Front. Psychol.11:1933. 10.3389/fpsyg.2020.01933

  • 66

    PowersA. R.MathysC.CorlettP. R. (2017). Pavlovian conditioning–induced hallucinations result from overweighting of perceptual priors.Science357596600. 10.1126/science.aan3458

  • 67

    RahnevD.FlemingS. M. (2019). How experimental procedures influence estimates of metacognitive ability.Neurosci. Conscious.2019:niz009. 10.1093/nc/niz009

  • 68

    RouaultM.DayanP.FlemingS. M. (2019). Forming global estimates of self-performance from local confidence.Nat. Commun.10111. 10.1038/s41467-019-09075-3

  • 69

    RouaultM.FlemingS. M. (2020). Formation of global self-beliefs in the human brain.Proc. Natl. Acad. Sci. U.S.A.1172726827276. 10.1073/pnas.2003094117

  • 70

    RouaultM.McWilliamsA.AllenM. G.FlemingS. M. (2018a). Human metacognition across domains: Insights from individual differences and neuroimaging.Personal. Neurosci.1113. 10.1017/pen.2018.16

  • 71

    RouaultM.SeowT.GillanC. M.FlemingS. M. (2018b). Psychiatric symptom dimensions are associated with dissociable shifts in metacognition but not task performance.Biol. Psychiatry84443451. 10.1016/j.biopsych.2017.12.017

  • 72

    RounisE.ManiscalcoB.RothwellJ. C.PassinghamR. E.LauH. (2010). Theta-burst transcranial magnetic stimulation to the prefrontal cortex impairs metacognitive visual awareness.Cogn. Neurosci.1165175. 10.1080/17588921003632529

  • 73

    RouyM.SaliouP.NalborczykL.PereiraM.RouxP.FaivreN. (2021). Systematic review and meta-analysis of the calibration of confidence judgments in individuals with schizophrenia spectrum disorders.Neurosci. Biobehav. Rev.126329337. 10.1016/j.neubiorev.2021.03.017

  • 74

    SadeghiS.EkhtiariH.BahramiB.AhmadabadiM. N. (2017). Metacognitive deficiency in a perceptual but not a memory task in methadone maintenance patients.Sci. Rep.718. 10.1038/s41598-017-06707-w

  • 75

    SellersR.VareseF.WellsA.MorrisonA. P. (2017). A meta-analysis of metacognitive beliefs as implicated in the self-regulatory executive function model in clinical psychosis.Schizophr. Res.1797584. 10.1016/j.schres.2016.09.032

  • 76

    SeowT. X. F.RouaultM.GillanC. M.FlemingS. M. (2021). How local and global metacognition shape mental health.Biol. Psychiatry90436446. 10.1016/j.biopsych.2021.05.013

  • 77

    TekcanA. I.TopcuogluV.KayaB. (2007). Memory and metamemory for semantic information in obsessive-compulsive disorder.Behav. Res. Ther.4521642172. 10.1016/j.brat.2006.10.002

  • 78

    VaccaroA. G.FlemingS. M. (2018). Thinking about thinking: A coordinate-based meta-analysis of neuroimaging studies of metacognitive judgements.Brain Neurosci. Adv.2114. 10.1177/2398212818810591

  • 79

    Van OosterhoutB.SmitF.KrabbendamL.CasteleinS.StaringA. B. P.Van Der GaagM. (2016). Metacognitive training for schizophrenia spectrum patients: A meta-analysis on outcome studies.Psychol. Med.464757. 10.1017/S0033291715001105

  • 80

    WilsonD. B. (2021). Wilson’s Meta-Analysis Page.Virginia: George Mason University. http://mason.gmu.edu/~dwilsonb/

  • 81

    WilsonD. B. (2001). Practical Meta-Analysis Effect Size Calculator [Online calculator]. Avilable online at: https://campbellcollaboration.org/research-resources/effect-size-calculator.html(accessed October 20th, 2022).

  • 82

    WrightA.NelsonB.FowlerD.GreenwoodK. (2020). Perceptual biases and metacognition and their association with anomalous self experiences in first episode psychosis.Conscious. Cogn.77118. 10.1016/j.concog.2019.102847

Summary

Keywords

metacognition, mental disorder, psychiatric, metacognitive sensitivity, experimental, decision making, systematic review and meta-analysis

Citation

Hohendorf M and Bauer M (2023) Metacognitive sensitivity and symptoms of mental disorder: A systematic review and meta-analysis. Front. Psychol. 14:991339. doi: 10.3389/fpsyg.2023.991339

Received

11 July 2022

Accepted

02 January 2023

Published

02 February 2023

Volume

14 - 2023

Edited by

Gerit Pfuhl, UiT The Arctic University of Norway, Norway

Reviewed by

Xiaohong Wan, Beijing Normal University, China; Jenifer L. Vohs, Indiana University School of Medicine, United States

Updates

Copyright

*Correspondence: Marianne Hohendorf, Markus Bauer,

This article was submitted to Cognition, a section of the journal Frontiers in Psychology

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics