MINI REVIEW article

Front. Behav. Neurosci., 07 April 2025

Sec. Pathological Conditions

Volume 19 - 2025 | https://doi.org/10.3389/fnbeh.2025.1543005

This article is part of the Research TopicCognitive impairments in schizophrenia, bipolar disorder, and major depression: Dissecting common and divergent featuresView all 3 articles

Current limitations in technology-based cognitive assessment for severe mental illnesses: a focus on feasibility, reliability, and ecological validity

  • Department of Mental and Physical Health and Preventive Medicine, University of Campania “Luigi Vanvitelli”, Naples, Italy

Cognitive impairments are frequently observed in subjects with severe mental illnesses (SMI), leading to a remarkable impact in their real-world functioning. Well-validated and gold standard instruments are available for the assessment of cognitive deficits, but different limitations should be considered, such as the need for specific training, lengthy administration times, practice effects, or reliance on subjective reports. Recent advances in digital technologies, such as ecological momentary assessments (EMA), virtual reality (VR), and passive digital phenotyping (DP), offer promising complementary approaches for capturing real-world cognitive functioning. In the current mini-review, we examine current research gaps that limit the application of these technologies, with a specific focus on feasibility, reliability and ecological validity. EMA may capture real-world functioning by increasing the number of evaluations throughout the day, but its use might be hindered by high participant burden and missing data. Furthermore, to achieve an accurate interpretation of EMA, studies should account for sampling and moment selection biases and the presence of several confounding factors. DP faces significant ethical and logistical challenges, including privacy and informed consent concerns, as well as challenges in data interpretation. VR could serve as a platform for both more ecologically valid cognitive assessments and rehabilitation interventions, but current barriers include technological and psychometric limitations, underdeveloped theoretical frameworks, and ethical considerations. Addressing these issues is crucial for ensuring that these novel technologies can effectively serve as valuable complements to traditional neuropsychological cognitive batteries.

1 Introduction

Cognitive deficits are a core feature of severe mental illnesses (SMI) and an important determinant of disability and impairments in real-world functioning (Sumiyoshi et al., 2019; Perrottelli et al., 2022; Vita et al., 2022a; Dragioti et al., 2023; Melillo et al., 2023; Mesholam-Gately et al., 2023; Firth et al., 2024; Giordano et al., 2024; Perrottelli et al., 2024), for which current therapies fail to provide satisfactory outcomes (Vita et al., 2018; Correll et al., 2023; Sampogna et al., 2023; Starzer et al., 2023). Among SMIs, cognitive deficits are particularly significant in schizophrenia, as they are often present even before the onset of the disorder, tend to persist during periods of clinical stability, and have a significant effect on everyday life functioning (Staal et al., 2000; Kravariti et al., 2007; Mesholam-Gately et al., 2009; Galderisi et al., 2020; Vita et al., 2022b; Handest et al., 2023). Given their clinical relevance, a reliable and feasible assessment of cognitive impairments is crucial for both clinical practice and research aimed at developing effective treatments (Hays et al., 2019; Vita et al., 2022b; Bucci et al., 2023; Melillo et al., 2024a).

Currently, assessment of cognitive functioning is mainly conducted either through performance-based cognitive tests or interviews, but both approaches have certain limitations. Pen-and-paper performance-based tests provide an objective method for evaluating cognitive deficits and gold-standard neuropsychological batteries are currently available and used in clinical practice to characterize patients’ cognitive profiles (Green et al., 2004; Yatham et al., 2010; Nuechterlein et al., 2023). However, despite their advantages, they still present different limitations. Firstly, they are susceptible to practice effects, becoming less effective in detecting changes in performance over time (Sahoo and Grover, 2022; Pezzella et al., 2024). Secondly, standardized cognitive batteries require specific training and have long administration times, thus limiting their effective implementation in routine clinical practice. Furthermore, traditional performance-based tests have been criticized for their lack of ecological validity, i.e., their ability to accurately capture and evaluate real-world cognitive functioning. Indeed, as these tests are typically conducted in controlled, clinical settings, they may not effectively capture how cognitive functioning translates to everyday life situations which involve distractions, multitasking demands, and emotional pressures (Sahoo and Grover, 2022). In addition, cognitive tests included in neuropsychological batteries are designed to isolate and assess cognitive domains individually. As a result, they do not resemble real-world cognitive challenges and tasks, which typically involve the complex interplay of multiple cognitive domains and are influenced by various factors, including, but not limited to, environmental distractors.

To address the challenges of extensive training requirements and lengthy administration times associated with traditional pen-and-paper cognitive batteries, computerized and tablet-based tests have been developed and adopted in recent decades. Computerized versions of pen-and-paper tests (Atkins et al., 2017), such as the tablet version of the Brief Assessment of Cognition in schizophrenia (BACS) and newly developed computerized batteries (Sahakian and Owen, 1992; Keefe et al., 2004; Levaux et al., 2007; Robbins et al., 2010) such as the Cambridge Neuropsychological Test Automated Battery (CANTAB) are relatively frequently used in clinical practice, providing scores comparable to pen-and-paper tests while requiring less supervision and clinical expertise (Hays et al., 2019). However, particularly when they are simple adaptations of pen-and-paper batteries, the introduction of technological versions of neuropsychological tests does not address the lack of ecological validity (Levaux et al., 2007; Feenstra et al., 2017; Harris et al., 2024). Furthermore, the absence of direct supervision and monitoring undermines the reliability of results, for instance introducing the risk of errors due to self-administration of tests (Levaux et al., 2007; Feenstra et al., 2017).

The evidence of the limited ecological validity and correlation with real-life functioning of both traditional and computerized tools prompted research efforts to develop new assessment methods. In relation to schizophrenia, within the Measurement and Treatment Research to Improve Cognition in Schizophrenia (MATRICS) initiative, the Food and Drug Administration indicated the need to integrate traditional cognitive batteries with co-primary outcomes focusing on the patients’ and caregivers’ perspectives on the real-life burden of cognitive impairment through interview-based evaluations like the cognitive assessment Interview (CAI) (Ventura et al., 2010; Cuesta et al., 2021; Bucci et al., 2023; Pezzella et al., 2024). Interview-based evaluations provide an index of how cognitive impairments impact everyday functioning in people with schizophrenia by collecting both patients’ and caregivers’ perspectives. Therefore, these assessments might detect subtle cognitive deficits that are subjectively experienced, but not captured by objective measures (Giordano et al., 2022a; Pezzella et al., 2024). Additionally, their administration requires less time and training, and their scores are not subject to practice effects, so they may more reliably detect improvements over short periods. However, as they rely on the self-reports of patients and caregivers, they can be influenced by psychopathology and biased by the individuals’ insight and capacity to assess potential cognitive difficulties (Petersen et al., 2019; Zimmerman, 2024). For instance, individuals with depressive symptoms tend to complain more about cognitive difficulties, while older subjects who experienced slow cognitive decline may be less aware of their difficulties and may have developed a series of compensation strategies in real-life (Petersen et al., 2019). Similarly, in schizophrenia, these tools can be affected by other psychopathological dimensions such as negative symptoms, which may bias both self-reported assessments and evaluations provided by caregivers (Handest et al., 2023; Rucci et al., 2023). Lastly, since interview-based assessments such as the CAI rely also on information provided by caregivers regarding the patient’s cognitive functioning, the validity of these scales is affected by variability in caregivers’ familiarity with the patient, inconsistent agreement between their evaluations and the patient’s self-assessment, and the frequent unavailability of informants (Bucci et al., 2023).

The highlighted limitations of both traditional and interview-based assessment have contributed to advancing research efforts for the development of complementary approaches. During the past two decades, recent advances in digital technologies, including ecological momentary assessments (EMA), virtual reality (VR), and passive digital phenotyping (DP), have emerged as promising tools to address the limitations of current assessment instruments. Technology-based cognitive assessment tools share the advantage that they do not require extensive training and time-consuming procedures for administration, making them more practical for both clinical and research settings. Additionally, their higher flexibility and adaptability to individual needs, as well as increasing accessibility, hold promise of delivering person-tailored assessments. Importantly, these tools have all been developed to enhance the ecological validity of cognitive assessments, either by delivering remote evaluations of real-life cognitive functioning, as in the case of EMA and DP, or by replicating real-life tasks and scenarios in virtual environments, as in the case of VR. These advantages position technology-based cognitive assessment tools as potential co-primary measures to traditional cognitive evaluations.

In the present paper, we provide a concise overview of the current state of the art of these approaches with the particular goal of highlighting current gaps currently limiting their implementation in clinical practice and research paradigms.

2 Methods

The present study is a mini-review aimed at synthesizing key findings on technology-based cognitive assessments in severe mental illness, identifying methodological considerations, feasibility aspects, and ethical challenges. For a more detailed description of the methodology, please refer to the Methods section of the Supplementary Material.

3 Results

3.1 Ecological momentary assessment

Ecological momentary assessment (EMA) is a method that provides real-time assessment of individuals through digital devices. EMA typically involves multiple assessments throughout the day, which take place in real-life settings and are initiated by the participant in response to alerts or push notifications at fixed or random intervals, depending on the designed paradigm (Parrish et al., 2021; Shvetz et al., 2021; Torous et al., 2021; Stone et al., 2023). By capturing data in real-world settings, EMA aims to enhance ecological validity and temporal resolution compared to traditional methods and to provide a more accurate representation of daily functioning (Smyth and Ebner-Priemer, 2025). Remote cognitive assessments using EMA via smartphones or smart watches are increasingly employed to evaluate cognition. Depending on their content, EMA paradigms of cognitive assessments can be categorized as either performance-based or interview-based. Performance-based assessments vary widely and include traditional cognitive tasks, adapted for EMA devices (Parrish et al., 2021; Shvetz et al., 2021), as well as tests specifically designed for EMA applications. Interview-based assessments, similar to traditional interviews evaluating functional capacity and cognitive performance, rely on subjective self-reports. The purported advantages of EMA, compared to traditional methods, lie in its ability to gain insights into the influence of naturalistic settings on cognitive functioning, thus delivering more ecologically valid evaluations of an individual’s cognitive functioning. In addition, EMA can achieve higher temporal resolution through multiple assessments per day, allowing it to capture intra-individual variations in cognitive performance through time (both circadian patterns and medium/long-term trajectories) (Moore et al., 2017). This approach has the potential to provide novel insights into the highly dynamic nature of cognitive functioning and the influence of various factors, including individual characteristics (e.g., mood, motivation), medication effects and environmental influences (e.g., noise or social distractors) (Moore et al., 2017; Weizenbaum et al., 2020).

However, these premises pose significant challenges. Firstly, the complex nature of EMA tools and the continuously evolving methodologies complicate the estimation and interpretation of their psychometric properties (Stone et al., 2023). Additionally, reliability estimates are not commonly reported in applied EMA research, and are limited to feasibility studies with low sample sizes (Liu et al., 2019; Moore et al., 2022; Singh et al., 2023; Stone et al., 2023; Harmon et al., 2024; Henneghan et al., 2025). Recent validation studies conducted in clinical and non-clinical populations have yielded varying results in relation to both within-person and between-person variability (Moore et al., 2022; Singh et al., 2023; Harmon et al., 2024; Henneghan et al., 2025). While lower within-person reliability—and thus higher intra individual variability— may be desirable in EMA assessment, given that its goal is to track the dynamic fluctuations in cognitive functioning, these observed changes must still reflect true and clinically meaningful variations (Stone et al., 2023). However, capturing real-world cognitive performance requires accounting for the numerous contextual variables that influence participants’ daily lives. This is necessary not only to interpret accurately the within-person variability of cognitive performances and to draw inferences on its potential causes but also to allow comparison between individuals, which is necessary for the development of normative data (Holmlund et al., 2019). Two possible examples are the environmental noise and external distractors. For instance, in the study by Hawks et al. (2023), momentary cognitive performance was worse when assessments were completed in environments perceived as noisy or busy; when participants reported difficulty concentrating before the assessment; and when participants experienced interruptions during the assessment. This is particularly problematic as a growing body of literature indicates that environmental features are often correlated with sociodemographic characteristics (e.g., noise is higher among communities with mid-to-low incomes per capita), further complicating the interpretation of results (Dreger et al., 2019; Huang et al., 2021; Hayward and Helbich, 2024). Many studies have attempted to account for environmental noise and other distractors by asking participants about their surroundings, but self-reports may be influenced by deficits in cognitive functioning. For instance, individuals with higher levels of cognitive impairments may describe their surroundings as noisier and more disruptive. Thus, subjective reports would probably need to be complemented by more objective evaluations, such as recordings of acoustic pollution or eye movement tracking, in order to detect distractions (Zamora et al., 2017; Yaroslavsky et al., 2019; Sweere et al., 2022; Hawks et al., 2023). These considerations suggest that EMA may require multifaceted data to produce evaluations that are truly ecologically valid (Weizenbaum et al., 2020).

Other current limitations of EMA depend on the involvement of multiple assessments per day to achieve higher temporal resolution and ecological validity. A first consequential challenge is related to practice effects when EMA employ performance-based tests (Hays et al., 2019), as repeated tasks may lead to improvements in performance over time. Recently, researchers have aimed to limit this effect by increasing variations in the designs and types of tasks presented to prevent the participant from learning task-specific strategies (Cohen et al., 2024). However, as highlighted by Stone et al. (2023), the most pressing challenge of multiple assessments is the amount of time and effort they require to be completed by the participants (Stone et al., 2023). These factors might influence the willingness of participants to participate, increasing the drop-out rates in research studies, and posing a risk of sampling bias in protocols using EMA. Subjects included in these studies may show differences in sociodemographic and cognitive profiles, as compared with patients who are either unwilling or unable to participate, as well as from those who fail to meet the minimum participation requirements (i.e., subjects who do not complete a minimum number of assessments during the day due to difficulties in utilizing digital devices) (Stone et al., 2023). In addition, the complexity of the EMA impacts the rate of compliance of participants (the ratio of answered prompts in relation to the total number of prompts scheduled) (Moore et al., 2017), due to the high time and effort demand of these multiple assessments (Eisele et al., 2022; Wrzus and Neubauer, 2023). Low compliance rates are of particular importance in EMA, because they pose the issue of how to interpret missing data, which may be the result of a combination of different factors. These include the presence of financial incentives for the patients, the number of prompts, the time interval between prompts, and the intelligence quotient (IQ), age and severity of symptoms of the patients. In addition, missing data may be the result of technical issues (e.g., internet outages, software updates), as well as of “moment selection bias”, i.e., data missed by contextual factors (e.g., participants may avoid EMA assessments in inappropriate or inconvenient situations) or, most importantly, factors directly related to the outcome itself (e.g., if the participant is asked to complete a cognitive task after periods of cognitive fatigue, or in a more distracting environment) (Wu et al., 2021; Stone et al., 2023; Reiter and Schoedel, 2024). This issue is particularly crucial, as data missing because of moment selection bias hinders the ecological validity and generalizability of the assessments (Stone et al., 2023). Lastly, alerts and push notifications for EMA can inadvertently draw attention from individuals near the participants, which could compromise privacy by revealing their participation in a research study. In fact, EMA, as other digital tools, present challenges related to privacy, data security and informed consent, which will be discussed in greater detail in the section “4 Discussion” (Galderisi et al., 2024).

3.2 Passive digital phenotyping

Passive DP is an emerging approach in psychiatric research that gathers real-time behavioral data via digital devices, without requiring the active participation of the assessed individual. The acquisition of information in passive DP can involve multiple devices, including not only smartphone sensors (such as Global Positioning Systems (GPS), accelerometers, keystroke dynamics, proximity detectors and ambient light sensors) but also a variety of other wearable devices (e.g., smart watches collecting heart rate variability, skin conductance, and blood pressure) (Khokhlov et al., 2020). DP is gaining traction as a tool for assessing cognitive function in various neuropsychiatric conditions (Cornet and Holden, 2018; Nguyen et al., 2023; Terhorst et al., 2024). Several parameters have been proposed as digital signatures associated with cognitive impairment, such as heart rate, phone usage, sleep data (Chen et al., 2019), keystroke patterns and typing speed (Dagum, 2018; Zulueta et al., 2018), GPS locations (Botros et al., 2022) and gait speed (Rasmussen et al., 2019). In a study by Zulueta et al. (2018) for instance, subjects with bipolar disorder in depressive states showed slower typing speeds and increased autocorrect usage, possibly reflecting impaired attention and self-monitoring, while participants in manic states showed a significant reduction in the use of the backspace key, possibly indicating decreased concentration.

Similarly to EMA, passive DP aims to achieve higher ecological validity by collecting data on cognitive performance in real-world environments. Importantly, it imposes little to no burden on participants, as data collection occurs in the background of the device without requiring the participant’s constant awareness. Thus, in comparison to EMA, DP cognitive assessments are not influenced by the participants’ adherence. In addition, they are less affected by the risks commonly associated with other unsupervised assessments (both computerized cognitive batteries and EMA), such as intentional manipulation of responses (i.e., “cheating”). Additionally, DP offers potentially unlimited temporal resolution, as it can involve the unobtrusive and continuous collection of large volumes of data, thus facilitating highly sensitive and precise assessments of intra-individual variations. This data allows researchers to dynamically adjust the resolution of their analyses — zooming out, to identify broader trends over weeks or months; zooming in, to explore moment-to-moment fluctuations in relation to specific contexts, times of day, or environmental triggers. This capability may eventually provide unprecedented insights into the temporal and spatial dynamics of cognitive functioning, enabling a more nuanced understanding of psychiatric phenomena.

Despite these significant opportunities, passive DP presents notable challenges. First, the literature on its reliability, validity, and convergence with test-based assessments is still in an early stage of development. Similarly to EMA, methods for estimating and documenting the reliability of passive DP have yet to be fully established. Furthermore, the clinical significance of many proposed digital biomarkers has not yet been thoroughly investigated (Cornet and Holden, 2018; Cohen et al., 2021; Terhorst et al., 2024). Indeed, available studies addressing the convergence between digital biomarkers and clinical ratings often included a low number of subjects, thus limiting the generalizability of their findings (Cornet and Holden, 2018). Several studies report limited or no relationship between passively sensed data and validation measures (Cornet and Holden, 2018; Cohen et al., 2021). These issues pose a significant challenge not only due to the scarcity of studies but also because of remarkable methodological difficulties (Hernandez et al., 2023; Hackett et al., 2024). Indeed, passive DP collects vast amounts of continuous, high-resolution data across diverse temporal and spatial contexts, whereas traditional neuropsychological assessments are typically cross-sectional and designed to assess specific cognitive domains. This discrepancy in resolution makes it difficult to determine the convergence between DP data and standard psychometric evaluations (Cohen et al., 2021). Moreover, similarly to EMA, DP of cognitive performance is challenged by the difficulty of accounting for the numerous contextual variables that influence participants’ daily lives, including environmental noise and external distractors (Hawks et al., 2023). In other words, DP shares with EMA the challenge of the interpretation of both the inter- and intra-individual variability of cognitive performances, and thus the estimation of the reliability of its assessments. The issue is particularly pressing for DP, given the higher risk of misinterpreting the collected digital biomarkers. For example, inactivity detected through GPS may be interpreted as a digital signature of avolition in subjects with schizophrenia, but could be wrongfully interpreted without considering factors like disability or socioeconomic barriers. Finally, issues linked to the feasibility of the use of DP also need to be addressed. In fact, given their strong reliance on the involved digital device, passive DP approaches are particularly susceptible to technical issues, including battery drainage and lack of sensor precision (Cornet and Holden, 2018). In addition, the sheer volume of data which can be collected through passive DP raises practical and logistical issues as well as ethical concerns. Indeed, questions regarding data storage, ownership, protection and access are particularly pressing, especially when collecting highly sensitive data, such as location, behavioral patterns, phone records. Moreover, as the scope of data collection of a study protocol may evolve over time, it becomes essential to ensure that informed consent processes are dynamic and adaptable (Holmlund et al., 2019). Participants must remain fully informed of how their data is being collected, analyzed, and potentially repurposed throughout the study. Addressing these challenges is essential to harness the full potential of passive DP while maintaining ethical and practical standards.

3.3 Virtual reality-based assessments

Virtual reality (VR) is a technology that allows users to immerse themselves in a three-dimensional computer-generated scenario (Chirico et al., 2016). VR systems are usually composed of a head-mounted 3D display (HMD), motion-tracking sensors, and controllers. The HMD immerses users in a virtual world, with motion-tracking sensors ensuring the sense of presence by replicating movements and controllers enabling interaction with virtual objects.

The last three decades have seen rapid growth in the application of VR across various clinical domains, including pain management (Rousseaux et al., 2020; Wittkopf et al., 2020; Melillo et al., 2022b; Melillo et al., 2024b), treatment of psychiatric conditions (particularly anxiety-related disorders) (Rus-Calafell et al., 2018; Emmelkamp et al., 2020; Riva et al., 2021; Wiebe et al., 2022), physical rehabilitation (Chirico et al., 2016; Groenveld et al., 2022; Melillo et al., 2022a; Sakdalan and Mitchell, 2023), and for medical and surgical training (Barré et al., 2019). VR can immerse participants in a digital environment, simulating real-world situations, while reducing unwanted sensory stimuli and minimizing many potential confounding variables (Fleming et al., 2023). Consequently, VR environments can also be employed to measure cognitive functioning in settings that are more closely aligned with real-life scenarios while still maintaining controlled and standardized settings (Harvey et al., 2022). In light of this potential, VR in SMI has been tested particularly for the assessment of functional capacity, by measuring performance in the execution of digital simulations of everyday life tasks (Ventura et al., 2020). In relation to this field, tools such as the Virtual Reality Functional Capacity Assessment Test (VRFCAT) have shown good test-retest reliability as well as good convergent validity with traditional measures of functional capacity (Keefe et al., 2016; Harvey et al., 2019). In relation to cognitive assessment, in the last decade, VR has been tested to evaluate specific cognitive domains by administering cognitive tasks embedded within the simulation of more complex real-life activities (Rus-Calafell et al., 2018; Miskowiak et al., 2021). For instance, verbal memory can be assessed by asking participants to recall shopping list items in a VR-replicated supermarket environment. It can also be very efficient to assess spatial memory, as individuals may be asked to remember the correct route to successfully navigate through three-dimensional scenarios (e.g., mazes) (Spieker et al., 2012; Wilkins et al., 2013). Additionally, it has been used to assess executive functioning through tests where participants perform tasks in dynamic environments requiring planning while ignoring distracting stimuli (e.g., running multiple errands in a shopping centre) (Dawson et al., 2009; Kirkham et al., 2024). More in general, VR is particularly interesting as it may help assessing the influence of specific cognitive deficits on real-world functioning, and provide simultaneous assessments of functional capacity and cognitive deficits (Ju et al., 2024). Several VR-based assessment tools have been developed, but the literature on their psychometric properties and particularly on their variability is very sparse (Rus-Calafell et al., 2018; Park, 2022; Kirkham et al., 2024). A recent systematic review on the reliability of VR measures of executive functioning showed that only 5 out of the 19 retrieved trials reported measures of internal consistency for the implemented tool, while none reported an index for test-retest reliability (Kirkham et al., 2024). Similarly, the convergent validity of VR tools, as compared to traditional cognitive tasks, remains to be fully established. A recent meta-analysis by Lee et al. (2024) examined the validity of VR-based tools, as compared to traditional neuropsychological tests for evaluating executive functions, both as a whole cognitive construct or with three subcomponents - cognitive flexibility, attention, and inhibition. The meta-analysis revealed statistically significant positive correlations between scores of VR-based assessments and traditional measures across the three different subcomponents of cognition considered (Lee et al., 2024). However, as highlighted by the authors, the studies presented a high heterogeneity in terms of the correlation effect sizes and showed evidence of publication bias in some of the studies included (Lee et al., 2024). This heterogeneity aligns with findings from studies on other cognitive domains, such as spatial and episodic memory, where research on the validity of VR approaches has also highlighted varying degrees of divergence between VR-based and traditional measures (Miskowiak et al., 2021; Jespersen et al., 2024; Mancuso et al., 2024). Additionally, VR assessment tools often exhibit ceiling effects, limiting their capacity to discriminate different levels of cognitive impairments (Miskowiak et al., 2021).

Beyond validity, a second challenge for VR-based tools concerns their feasibility and acceptability, due to the technological expertise required to use VR systems. This issue may pose a significant barrier for the clinical application of VR for subjects with SMI with cognitive impairment or limited digital proficiency, as it is often the case for people affected by psychiatric disorders, raising potential ethical concerns regarding equity in its use (Garrett et al., 2018). Furthermore, the identification of potentials risks associated to the use of VR in specific categories of subjects with SMI is crucial. Particular attention should be given to the potential long-term effects of VR exposure, such as an increased risk of dissociative symptoms, including derealization and depersonalization (Madary and Metzinger, 2016). To address these risks, the establishment of empirically-driven exclusion criteria is essential, including thresholds on dissociative experience scales to identify individuals who may be particularly vulnerable. Additionally, these factors limit the application of VR for remote assessments, and in clinical trials these are often performed under supervision or following proper training of the participants.

Finally, VR implementation in clinical and research settings is far from achieving consistent integration. As the field is experiencing rapid developments of both the hardware and software components, studies even a year apart are often difficult to compare. This issue is complicated by the fact that often scientific papers do not provide technological specifics, using the term VR to define significantly different technologies (e.g., non-immersive/computer-based and immersive VR), and rarely address theoretical aspects or providing robust theoretical explanations of how VR constructs (e.g., immersivity, tele-presence) apply to the specific areas under investigation (Atkins et al., 2015; Negu? et al., 2016; Garrett et al., 2018; Howard, 2019; Wiener et al., 2020).

Overall, as regard to cognitive assessments, significant challenges still remain in the use of VR, such as creating paradigms that effectively balance precision, ecological validity, and experimental control, while remaining adaptable and applicable to diverse populations and research needs (Krohn et al., 2020).

4 Discussion

Digital technologies represent a great opportunity to improve the assessment of cognitive functioning in individuals with SMI, thereby advancing research into their underlying mechanisms and potential treatment options. In the present paper, we aimed to offer an overview of the current research gaps, particularly on issues and possible limitations related to the feasibility, reliability, and validity of these new tools, with the aim of guiding future research and supporting their integration into both clinical practice and research paradigms. The growing interest in these approaches is further driven by their relevance to the longstanding debate between experimental control and ecological validity (Banaji and Crowder, 1989; Conway, 1991; Neisser and Hyman, 2000; Chaytor and Schmitter-Edgecombe, 2003). Indeed, cognitive assessment tools for SMI, traditionally, are designed to maintain high control of the experimental and environmental settings to reduce confounding factors. This approach aims to enhance the reliability and interpretability of the results, providing an accurate reflection of an individual’s cognitive abilities and facilitating the identification of neurobiological mechanisms and potential pharmacological targets (Green et al., 2004). While these methods are generally highly reliable and have good internal validity, they inherently lack ecological validity, as they fail to capture an individual’s true cognitive performance in real-world settings. To address this gap, recently, the FDA has recommended integrating performance-based measures with interview-based approaches capable of gathering insights regarding an individual’s perceived challenges in completing tasks linked to specific cognitive processes (Giordano et al., 2022b). Aligned with this view, differently from traditional cognitive batteries, newer digital approaches aim to enhance ecological validity by capturing cognitive functioning in more naturalistic settings. In relation to EMA and DP, these technologies carry the opportunity of remotely capturing novel data on cognitive functioning in real-world contexts with high temporal resolution. However, accurately assessing naturalistic cognitive performance requires accounting for the near-infinite contextual variables that may influence an individual’s assessment, which in turn are typically controlled in supervised, in-person settings. This is necessary for accurately interpreting both intra-individual and inter-individual variability, as well as their potential causes, which is essential for developing normative data (Holmlund et al., 2019).

In relation to EMA, further research is needed to address remarkable challenges such as participant burden, missing data, sampling biases, moment selection, and the psychometric properties of EMA strategies (Miskowiak et al., 2021; Stone et al., 2023; Lane et al., 2024). In relation to DP, the unique advantage of this methodology is the little to no burden on participants, as data collection occurs in the background of the device without requiring active engagement. This advantage is particularly relevant for schizophrenia since the reduced motivational drive intrinsic to the disorder can affect cognitive performance, highlighting the need for highly engaging yet low-demand tools for both assessment and intervention (Fervaha et al., 2014). Therefore, unlike EMA, DP assessments are less influenced by participant adherence and are not prone to risks such as intentional manipulation of responses. Moreover, DP provides potentially unlimited temporal resolution, enabling unobtrusive, continuous data collection that supports highly sensitive assessments of intra-individual variations. However, the literature on the reliability, validity and convergence with test-based assessments of passive DP cognitive assessments is still in its early stages. Additionally, feasibility issues, including technical challenges and concerns related to data storage, ownership, and protection, need to be addressed in passive DP approaches.

Another key advantage of these new technologies is their ability to operate across virtually unlimited geographical ranges, offering an accessible, highly customizable psychometric method for individuals who might otherwise face significant barriers to traditional assessments. Additionally, although their development may require higher initial financial investments, these technologies could reduce long-term costs by minimizing the need for in-person supervision and thus leveraging scalability. Indeed, evidence on internet- and mobile-based interventions, especially in schizophrenia, highlights their potential to lower healthcare costs, by decreasing the time required from therapists and the need for patient travel, despite higher upfront costs (Kählke et al., 2022). While data on the cost-effectiveness of digital cognitive assessments is limited, their scalability and accessibility suggest they may have a similar cost-saving potential (Griffiths et al., 2006; Hedman et al., 2012; Gómez Bergin and Craven, 2023).

One method that could potentially achieve a balance between the two guiding principles of ecological validity and controlled experimental settings is VR (Parsons, 2015). With the foreseeable future technological advancements, VR has the potential to achieve increasingly higher levels of ecological validity by creating digital scenarios that will be progressively more similar to the ones encountered in real life. At the same time, it allows for strict control over perceptual stimuli and assessment conditions (e.g., number and intensity of stimuli simulating real-world distractors), thus maintaining experimental control and reliability of evaluations. Importantly, VR technologies allow for the assessment of both cognitive functioning and functional capacity, helping to bridge the gap between these two aspects. Furthermore, they offer a platform for both assessment and treatment strategies, thus supporting a more direct translation of cognitive assessments into clinical interventions while ensuring more personalized, feasible, and reliable therapeutic frameworks (Emsley, 2023; Keshavan and Eack, 2023; Schäfer et al., 2023; Sampogna et al., 2024). However, the clinical application of VR paradigms in cognitive assessment will require further study regarding their psychometric properties, tool-specific limitations (e.g., motion sickness) and the thorough development of an underlying theoretical and ethical framework.

The adoption of these technologies raises important ethical considerations, including concerns about privacy and data security. Additionally, there is a need for careful informed consent processes, particularly in the case of extensive data collection methods – such as passive DP – which often have multiple or evolving purposes and destinations. Therefore, it is crucial to establish informed consent processes that are flexible and adaptable to these changes (Couper and Singer, 2013; Kunz et al., 2020; Nica, 2024; Sansone et al., 2024). Keeping in mind the target populations in SMI, balancing autonomy with parental or caregiver involvement remains critical to ensure ethical standards (Kourtis et al., 2019; Lakhtakia et al., 2022; Bartlett, 2024); this aspect is particularly complex in schizophrenia, where individuals affected by the disorder are more likely to lack the capacity to provide informed consent, as compared to healthy control groups, particularly if they present cognitive deficits, as supported by findings from recent meta-analyses (Jeste et al., 2006; Wang et al., 2017; Parsons, 2021). In order to help patients and their legal guardians better understand both the procedures and the potential risks of digital assessment, research could implement interactive tools to assess comprehension (e.g., VR-based demonstrations, digital follow-up questions), ensuring that participants provide fully informed consent. Furthermore, in the case of passive DP approaches, these tools also raise significant concerns about data privacy, protection and storage, as they collect detailed information such as geolocation and keystroke dynamics, potentially exposing sensitive personal details without participants fully understanding how their data will be used (Martinez-Martin et al., 2018; D’Alfonso et al., 2024; Martinez-Martin, 2024; Sansone et al., 2024; Tyano, 2024).

In addition to the previously discussed ethical, VR paradigms introduce tool-specific ethical concerns, including the representation of reality within VR, the autonomy of users, and the potential for unintended effects not only in the virtual setting (in virtuo), but also in real life (in vivo) (Madary and Metzinger, 2016). For instance, in virtual settings, the anonymity of avatars and lack of accountability can foster harmful behaviors, such as cyber harassment and cyberbullying. These issues, amplified by the immersive nature of VR, may provoke significant emotional distress (Kourtesis, 2024). Additionally, unintended effects of virtual-self representation may extend to real life (in vivo), as identification with virtual avatars has been shown to influence real-world behavior, promoting actions consistent with the avatar’s role, whether prosocial or antisocial (Rosenberg et al., 2013). These risks are amplified when VR interventions are used with at-risk groups such as children, the elderly, and individuals with SMI) (Madary and Metzinger, 2016).

This mini-review aimed to provide a brief explorative overview of the potential limitations of current technology-based cognitive assessment, with a particular focus on their feasibility, reliability, and ecological validity. However, some limitations should be noted. First, as mini-reviews prioritize conciseness over comprehensiveness, their scope is necessarily limited, so it may result in the exclusion of relevant studies or perspectives. In particular, we could not discuss all the specific technical issues, implementation requirements and regulatory and legal aspects of the discussed technologies. Furthermore, we decided not to discuss thoroughly data quality control methods and data standardization strategies of the discussed assessment tools. The second limitation is relevant to the narrative review methodology: as our work did not involve an exhaustive systematic search strategy, it may have potentially excluded some relevant studies. Furthermore, our work did not involve either a quantitative analysis of the retrieved evidence or a systematic assessment of the quality of included studies, as these were beyond the scope of our review. However, this work might serve as a basis for future quantitative systematic reviews and meta-analyses addressing these research questions and evaluating the robustness and quality of the associated evidence.

5 Conclusion and future perspectives

Technology-based assessments can provide valuable insights into cognitive impairment and serve as a complementary tool to traditional cognitive evaluations. By bridging the gap between cognitive performance measured in controlled experimental settings and real-world cognitive functioning, these technologies can significantly enhance our understanding of how cognitive impairments impact functional outcomes. This, in turn, may crucially contribute to the development of more effective and personalized cognitive and functional rehabilitation programs, ultimately addressing the disability burden associated with severe mental illnesses (Palumbo et al., 2022; Giuliani et al., 2024; Perna et al., 2024).

Addressing the challenges discussed in the present paper is therefore essential to ensure that these emerging technologies achieve these goals.

In relation to EMA and passive DP, a promising path forward may lie in the combined application of these tools. A successful example is the collection of passive DP data during EMA assessments (i.e., “paradata”). Paradata, as applied for this context, refers to auxiliary data collected during the administration of digital assessments (e.g., response times, navigation patterns, position). This approach may be particularly promising, as it may contribute to the development of more “context-aware” EMA assessments, by using digital devices to capture observations of both the individual and its environment (Burns et al., 2011). These data can help researchers infer aspects of cognitive performance or identify factors affecting data quality and validity. In addition, as DP paradata can provide data beyond the primary scope of the EMA assessment, it can provide more multidimensional observations of the participant. For example, researchers can analyze behavioral metrics like reaction times during survey responses to infer cognitive performance, even if the primary focus of the survey is on other dimensions, which may be of interest for the evaluation of cognitive functioning (e.g., questions evaluating mood or anxiety) (Dagum, 2018; McClain et al., 2019; Abi-Dargham et al., 2023; Berk, 2023; Hernandez et al., 2023; Tamminga, 2023; Voineskos, 2023). Thus, the combination of passive DP and EMA can improve the ecological validity of the assessments by providing context-aware and multidimensional observations. Given the high dependence on the precision of the employed applied digital devices, further studies should contribute to the development of standardized data quality control methods and algorithms (Cornet and Holden, 2018). In relation to VR, the development of robust theoretical frameworks for VR constructs will be critical for the study of the psychometric properties of VR-based evaluations. In addition, more detailed reporting of technological specifications is essential to enhance the comparison of study results and to support efforts toward standardization. Overall, further studies are needed to broaden the evidence base on the validity, feasibility and psychometric properties of these evaluation tools. Particularly in the case of EMA and DP, it is urgent to develop and implement improved reliability estimation methods (Stone et al., 2023). Future studies should always report relevant reliability and variability coefficients in order to enhance transparency and to inform future study design decisions. Future implementations should focus on improving accessibility for patients with varying levels of digital literacy. For instance, providing pre-configured devices or step-by-step guidance for patients and caregivers could help ensure more equitable access to these technologies. Importantly, there is a lack of a unified ethical and regulatory framework for the implementation of digital technologies in mental health care (Steindl, 2023; Galderisi et al., 2024). This is a crucial issue particularly in relation to EMA and DP, as the development of standardized frameworks for secure data handling and storage is a necessary step for their safe and ethical implementation.

Finally, as current studies are predominantly conducted in high-income countries, further research will need to involve cross-cultural validation and investigate the adaptability and feasibility of these tools in varied socio-economic and cultural contexts (Moore et al., 2022; Hernandez et al., 2023; Hackett et al., 2024; Harmon et al., 2024; Henneghan et al., 2025).

Author contributions

EC: Writing – original draft, Writing – review and editing. AM: Writing – original draft, Writing – review and editing. AP: Writing – original draft, Writing – review and editing. LG: Writing – review and editing. FM: Writing – review and editing. PP: Writing – review and editing. GG: Writing – original draft, Writing – review and editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnbeh.2025.1543005/full#supplementary-material

References

Abi-Dargham, A., Moeller, S. J., Ali, F., DeLorenzo, C., Domschke, K., Horga, G., et al. (2023). Candidate biomarkers in psychiatric disorders: State of the field. World Psychiatry 22, 236–262. doi: 10.1002/wps.21078

PubMed Abstract | Crossref Full Text | Google Scholar

Atkins, A. S., Stroescu, I., Spagnola, N. B., Davis, V. G., Patterson, T. D., Narasimhan, M., et al. (2015). Assessment of age-related differences in functional capacity using the virtual reality functional capacity assessment tool (VRFCAT). J. Prev. Alzheimers Dis. 2, 121–127. doi: 10.14283/jpad.2015.61

PubMed Abstract | Crossref Full Text | Google Scholar

Atkins, A. S., Tseng, T., Vaughan, A., Twamley, E. W., Harvey, P., Patterson, T., et al. (2017). Validation of the tablet-administered brief assessment of cognition (BAC App). Schizophr. Res. 181, 100–106. doi: 10.1016/j.schres.2016.10.010

PubMed Abstract | Crossref Full Text | Google Scholar

Banaji, M. R., and Crowder, R. G. (1989). The bankruptcy of everyday memory. Am. Psychol. 44, 1185–1193. doi: 10.1037/0003-066X.44.9.1185

Crossref Full Text | Google Scholar

Barré, J., Michelet, D., Truchot, J., Jolivet, E., Recanzone, T., Stiti, S., et al. (2019). Virtual reality single-port sleeve gastrectomy training decreases physical and mental workload in novice surgeons: An exploratory study. Obes. Surg. 29, 1309–1316. doi: 10.1007/s11695-018-03680-9

PubMed Abstract | Crossref Full Text | Google Scholar

Bartlett, P. (2024). Ethical challenges in mental health care: Moving beyond aspirations. World Psychiatry 23, 387–388. doi: 10.1002/wps.21231

PubMed Abstract | Crossref Full Text | Google Scholar

Berk, M. (2023). Biomarkers in psychiatric disorders: Status quo, impediments and facilitators. World Psychiatry 22, 174–176. doi: 10.1002/wps.21071

PubMed Abstract | Crossref Full Text | Google Scholar

Botros, A. A., Schuetz, N., Röcke, C., Weibel, R., Martin, M., Müri, R. M., et al. (2022). Eigenbehaviour as an indicator of cognitive abilities. Sensors 22:2769. doi: 10.3390/s22072769

PubMed Abstract | Crossref Full Text | Google Scholar

Bucci, P., Mucci, A., Giordano, G. M., Caporusso, E., Giuliani, L., Gibertoni, D., et al. (2023). Insight in cognitive impairment assessed with the cognitive assessment interview in a large sample of patients with schizophrenia. Eur. Arch. Psychiatry Clin. Neurosci. 274, 1343–1354. doi: 10.1007/s00406-023-01641-7

PubMed Abstract | Crossref Full Text | Google Scholar

Burns, M. N., Begale, M., Duffecy, J., Gergle, D., Karr, C. J., Giangrande, E., et al. (2011). Harnessing context sensing to develop a mobile intervention for depression. J. Med. Internet Res. 13:e55. doi: 10.2196/jmir.1838

PubMed Abstract | Crossref Full Text | Google Scholar

Chaytor, N., and Schmitter-Edgecombe, M. (2003). The ecological validity of neuropsychological tests: A review of the literature on everyday cognitive skills. Neuropsychol. Rev. 13, 181–197. doi: 10.1023/b:nerv.0000009483.91468.fb

PubMed Abstract | Crossref Full Text | Google Scholar

Chen, R., Jankovic, F., Marinsek, N., Foschini, L., Kourtis, L., Signorini, A., et al. (2019). “Developing measures of cognitive impairment in the real world from consumer-grade multimodal sensor streams,” in Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, (Anchorage, AK: Association for Computing Machinery).

Google Scholar

Chirico, A., Lucidi, F., De Laurentiis, M., Milanese, C., Napoli, A., and Giordano, A. (2016). Virtual reality in health system: Beyond entertainment. A mini-review on the efficacy of VR during cancer treatment. J. Cell Physiol. 231, 275–287. doi: 10.1002/jcp.25117

PubMed Abstract | Crossref Full Text | Google Scholar

Cohen, A. S., Schwartz, E., Le, T. P., Cowan, T., Kirkpatrick, B., Raugh, I. M., et al. (2021). Digital phenotyping of negative symptoms: The relationship to clinician ratings. Schizophr. Bull. 47, 44–53. doi: 10.1093/schbul/sbaa065

PubMed Abstract | Crossref Full Text | Google Scholar

Cohen, A., Joshi, D., Bondre, A., Chand, P. K., Chaturvedi, N., Choudhary, S., et al. (2024). Digital phenotyping correlates of mobile cognitive measures in schizophrenia: A multisite global mental health feasibility trial. PLoS Digit. Health 3:e0000526. doi: 10.1371/journal.pdig.0000526

PubMed Abstract | Crossref Full Text | Google Scholar

Conway, M. A. (1991). In defense of everyday memory. Am. Psychol. 46, 19–26. doi: 10.1037/0003-066X.46.1.19

Crossref Full Text | Google Scholar

Cornet, V. P., and Holden, R. J. (2018). Systematic review of smartphone-based passive sensing for health and wellbeing. J. Biomed. Inform. 77, 120–132. doi: 10.1016/j.jbi.2017.12.008

PubMed Abstract | Crossref Full Text | Google Scholar

Correll, C. U., Solmi, M., Cortese, S., Fava, M., Højlund, M., Kraemer, H. C., et al. (2023). The future of psychopharmacology: A critical appraisal of ongoing phase 2/3 trials, and of some current trends aiming to de-risk trial programmes of novel agents. World Psychiatry 22, 48–74. doi: 10.1002/wps.21056

PubMed Abstract | Crossref Full Text | Google Scholar

Couper, M. P., and Singer, E. (2013). Informed consent for web paradata use. Surv. Res. Methods 7, 57–67. doi: 10.18148/srm/2013.v7i1.5138

Crossref Full Text | Google Scholar

Cuesta, M. J., Sánchez-Torres, A. M., Lorente-Omeñaca, R., Moreno-Izco, L., and Peralta, V. (2021). Cognitive, community functioning and clinical correlates of the Clinical assessment interview for negative symptoms (CAINS) in psychotic disorders. Eur. Arch. Psychiatry Clin. Neurosci. 271, 1537–1546. doi: 10.1007/s00406-020-01188-x

PubMed Abstract | Crossref Full Text | Google Scholar

D’Alfonso, S., Coghlan, S., Schmidt, S., and Mangelsdorf, S. (2024). Ethical dimensions of digital phenotyping within the context of mental healthcare. J. Technol. Behav. Sci. 1–16. doi: 10.1007/s41347-024-00423-9

Crossref Full Text | Google Scholar

Dagum, P. (2018). Digital biomarkers of cognitive function. npj Digital Med. 1:10. doi: 10.1038/s41746-018-0018-4

PubMed Abstract | Crossref Full Text | Google Scholar

Dawson, D. R., Anderson, N. D., Burgess, P., Cooper, E., Krpan, K. M., and Stuss, D. T. (2009). Further development of the Multiple errands test: Standardized Scoring, reliability, and ecological validity for the Baycrest version. Arch. Phys. Med. Rehabil. 90, S41–S51. doi: 10.1016/j.apmr.2009.07.012

PubMed Abstract | Crossref Full Text | Google Scholar

Dragioti, E., Radua, J., Solmi, M., Gosling, C. J., Oliver, D., Lascialfari, F., et al. (2023). Impact of mental disorders on clinical outcomes of physical diseases: An umbrella review assessing population attributable fraction and generalized impact fraction. World Psychiatry 22, 86–104. doi: 10.1002/wps.21068

PubMed Abstract | Crossref Full Text | Google Scholar

Dreger, S., Schüle, S. A., Hilz, L. K., and Bolte, G. (2019). Social inequalities in environmental noise exposure: A review of evidence in the WHO European region. Int. J. Environ. Res. Public Health 16:1011. doi: 10.3390/ijerph16061011

PubMed Abstract | Crossref Full Text | Google Scholar

Eisele, G., Vachon, H., Lafit, G., Kuppens, P., Houben, M., Myin-Germeys, I., et al. (2022). The effects of sampling frequency and questionnaire length on perceived burden, compliance, and careless responding in experience sampling data in a student population. Assessment 29, 136–151. doi: 10.1177/1073191120957102

PubMed Abstract | Crossref Full Text | Google Scholar

Emmelkamp, P. M. G., Meyerbröker, K., and Morina, N. (2020). Virtual reality therapy in social anxiety disorder. Curr. Psychiatry Rep. 22:32. doi: 10.1007/s11920-020-01156-1

PubMed Abstract | Crossref Full Text | Google Scholar

Emsley, R. (2023). The future of psychopharmacology: Challenges beyond efficacy and tolerability. World Psychiatry 22, 82–83. doi: 10.1002/wps.21065

PubMed Abstract | Crossref Full Text | Google Scholar

Feenstra, H. E., Vermeulen, I. E., Murre, J. M., and Schagen, S. B. (2017). Online cognition: Factors facilitating reliable online neuropsychological test results. Clin. Neuropsychol. 31, 59–84. doi: 10.1080/13854046.2016.1190405

PubMed Abstract | Crossref Full Text | Google Scholar

Fervaha, G., Zakzanis, K. K., Foussias, G., Graff-Guerrero, A., Agid, O., and Remington, G. (2014). Motivational deficits and cognitive test performance in schizophrenia. JAMA Psychiatry 71, 1058–1065. doi: 10.1001/jamapsychiatry.2014.1105

PubMed Abstract | Crossref Full Text | Google Scholar

Firth, J., Torous, J., López-Gil, J. F., Linardon, J., Milton, A., Lambert, J., et al. (2024). From “online brains” to “online lives”: Understanding the individualized impacts of Internet use across psychological, cognitive and social dimensions. World Psychiatry 23, 176–190. doi: 10.1002/wps.21188

PubMed Abstract | Crossref Full Text | Google Scholar

Fleming, T., Poppelaars, M., and Thabrew, H. (2023). The role of gamification in digital mental health. World Psychiatry 22, 46–47. doi: 10.1002/wps.21041

PubMed Abstract | Crossref Full Text | Google Scholar

Galderisi, S., Appelbaum, P. S., Gill, N., Gooding, P., Herrman, H., Melillo, A., et al. (2024). Ethical challenges in contemporary psychiatry: An overview and an appraisal of possible strategies and research needs. World Psychiatry 23, 364–386. doi: 10.1002/wps.21230

PubMed Abstract | Crossref Full Text | Google Scholar

Galderisi, S., Rucci, P., Mucci, A., Rossi, A., Rocca, P., Bertolino, A., et al. (2020). The interplay among psychopathology, personal resources, context-related factors and real-life functioning in schizophrenia: Stability in relationships after 4 years and differences in network structure between recovered and non-recovered patients. World Psychiatry 19, 81–91. doi: 10.1002/wps.20700

PubMed Abstract | Crossref Full Text | Google Scholar

Garrett, B., Taverner, T., Gromala, D., Tao, G., Cordingley, E., and Sun, C. (2018). Virtual reality clinical research: Promises and challenges. JMIR Serious Games 6:e10839. doi: 10.2196/10839

PubMed Abstract | Crossref Full Text | Google Scholar

Giordano, G. M., Brando, F., Pezzella, P., De Angelis, M., Mucci, A., and Galderisi, S. (2022a). Factors influencing the outcome of integrated therapy approach in schizophrenia: A narrative review of the literature. Front. Psychiatry 13:970210. doi: 10.3389/fpsyt.2022.970210

PubMed Abstract | Crossref Full Text | Google Scholar

Giordano, G. M., Palumbo, D., Mucci, A., Ventura, J., Giuliani, L., Perrottelli, A., et al. (2022b). The Cognitive assessment interview (CAI): Association with neuropsychological scores and real-life functioning in a large sample of Italian subjects with schizophrenia. Schizophr. Res. 241, 161–170. doi: 10.1016/j.schres.2022.01.029

PubMed Abstract | Crossref Full Text | Google Scholar

Giordano, G. M., Sanmarchi, F., Mucci, A., Rucci, P., Brando, F., Caporusso, E., et al. (2024). External validation of the five domains of negative symptoms: Focus on cognition, functional capacity, and real-world functioning. Eur. Psychiatry 67:e3. doi: 10.1192/j.eurpsy.2023.2478

PubMed Abstract | Crossref Full Text | Google Scholar

Giuliani, L., Pezzella, P., Mucci, A., Palumbo, D., Caporusso, E., Piegari, G., et al. (2024). Effectiveness of a social cognition remediation intervention for patients with schizophrenia: A randomized-controlled study. Ann. General Psychiatry 23:52. doi: 10.1186/s12991-024-00535-9

PubMed Abstract | Crossref Full Text | Google Scholar

Gómez Bergin, A. D., and Craven, M. P. (2023). Virtual, augmented, mixed, and extended reality interventions in healthcare: A systematic review of health economic evaluations and cost-effectiveness. BMC Digital Health 1:53. doi: 10.1186/s44247-023-00054-9

Crossref Full Text | Google Scholar

Green, M. F., Nuechterlein, K. H., Gold, J. M., Barch, D. M., Cohen, J., Essock, S., et al. (2004). Approaching a consensus cognitive battery for clinical trials in schizophrenia: The NIMH-MATRICS conference to select cognitive domains and test criteria. Biol. Psychiatry 56, 301–307. doi: 10.1016/j.biopsych.2004.06.023

PubMed Abstract | Crossref Full Text | Google Scholar

Griffiths, F., Lindenmeyer, A., Powell, J., Lowe, P., and Thorogood, M. (2006). Why are health care interventions delivered over the internet? A systematic review of the published literature. J. Med. Internet Res. 8:e10. doi: 10.2196/jmir.8.2.e10

PubMed Abstract | Crossref Full Text | Google Scholar

Groenveld, T., Achttien, R., Smits, M., de Vries, M., van Heerde, R., Staal, B., et al. (2022). Feasibility of virtual reality exercises at home for post-COVID-19 condition: Cohort study. JMIR Rehabil. Assist. Technol. 9:e36836. doi: 10.2196/36836

PubMed Abstract | Crossref Full Text | Google Scholar

Hackett, K., Xu, S., McKniff, M., Paglia, L., Barnett, I., and Giovannetti, T. (2024). Mobility-based smartphone digital phenotypes for unobtrusively capturing everyday cognition, mood, and community life-space in older adults: Feasibility, acceptability, and preliminary validity study. JMIR Hum. Factors 11:e59974. doi: 10.2196/59974

PubMed Abstract | Crossref Full Text | Google Scholar

Handest, R., Molstrom, I. M., Gram Henriksen, M., Hjorthøj, C., and Nordgaard, J. (2023). A systematic review and meta-analysis of the association between psychopathology and social functioning in schizophrenia. Schizophr. Bull. 49, 1470–1485. doi: 10.1093/schbul/sbad075

PubMed Abstract | Crossref Full Text | Google Scholar

Harmon, S., Kocum, C. G., Ranum, R. M., Hermann, G., Farias, S. T., and Kiselica, A. M. (2024). The mobile everyday cognition scale (mECog): Development and pilot testing. Clin. Neuropsychol. 39, 451–470. doi: 10.1080/13854046.2024.2383333

PubMed Abstract | Crossref Full Text | Google Scholar

Harris, C., Tang, Y., Birnbaum, E., Cherian, C., Mendhe, D., and Chen, M. H. (2024). Digital neuropsychology beyond computerized cognitive assessment: Applications of novel digital technologies. Arch. Clin. Neuropsychol. 39, 290–304. doi: 10.1093/arclin/acae016

Crossref Full Text | Google Scholar

Harvey, P. D., Depp, C. A., Rizzo, A. A., Strauss, G. P., Spelber, D., Carpenter, L. L., et al. (2022). Technology and mental health: State of the art for assessment and treatment. Am. J. Psychiatry 179, 897–914. doi: 10.1176/appi.ajp.21121254

PubMed Abstract | Crossref Full Text | Google Scholar

Harvey, P. D., Khan, A., Atkins, A., and Keefe, R. S. (2019). Virtual reality assessment of functional capacity in people with Schizophrenia: Associations with reduced emotional experience and prediction of functional outcomes. Psychiatry Res. 277, 58–63. doi: 10.1016/j.psychres.2019.01.045

PubMed Abstract | Crossref Full Text | Google Scholar

Hawks, Z. W., Strong, R., Jung, L., Beck, E. D., Passell, E. J., Grinspoon, E., et al. (2023). Accurate prediction of momentary cognition from intensive longitudinal data. Biol. Psychiatry Cogn. Neurosci. Neuroimaging 8, 841–851. doi: 10.1016/j.bpsc.2022.12.002

PubMed Abstract | Crossref Full Text | Google Scholar

Hays, R., Henson, P., Wisniewski, H., Hendel, V., Vaidyam, A., and Torous, J. (2019). Assessing cognition outside of the clinic: Smartphones and sensors for cognitive assessment across diverse psychiatric disorders. Psychiatr. Clin. North Am. 42, 611–625. doi: 10.1016/j.psc.2019.08.003

PubMed Abstract | Crossref Full Text | Google Scholar

Hayward, M., and Helbich, M. (2024). Environmental noise is positively associated with socioeconomically less privileged neighborhoods in the Netherlands. Environ. Res. 248:118294. doi: 10.1016/j.envres.2024.118294

PubMed Abstract | Crossref Full Text | Google Scholar

Hedman, E., Ljótsson, B., and Lindefors, N. (2012). Cognitive behavior therapy via the Internet: A systematic review of applications, clinical efficacy and cost-effectiveness. Expert. Rev. Pharmacoecon Outcomes Res. 12, 745–764. doi: 10.1586/erp.12.67

PubMed Abstract | Crossref Full Text | Google Scholar

Henneghan, A. M., Paolillo, E. W., Van Dyk, K. M., Franco-Rocha, O. Y., Bang, S., Tasker, R., et al. (2025). Feasibility and psychometric quality of smartphone administered cognitive ecological momentary assessments in women with metastatic breast cancer. Digit. Health 11:20552076241310474. doi: 10.1177/20552076241310474

PubMed Abstract | Crossref Full Text | Google Scholar

Hernandez, R., Hoogendoorn, C., Gonzalez, J. S., Jin, H., Pyatak, E. A., Spruijt-Metz, D., et al. (2023). Reliability and validity of noncognitive ecological momentary assessment survey response times as an indicator of cognitive processing speed in people’s natural environment: Intensive longitudinal study. JMIR Mhealth Uhealth 11:e45203. doi: 10.2196/45203

PubMed Abstract | Crossref Full Text | Google Scholar

Holmlund, T. B., Foltz, P. W., Cohen, A. S., Johansen, H. D., Sigurdsen, R., Fugelli, P., et al. (2019). Moving psychological assessment out of the controlled laboratory setting: Practical challenges. Psychol. Assess. 31, 292–303. doi: 10.1037/pas0000647

PubMed Abstract | Crossref Full Text | Google Scholar

Howard, M. C. (2019). Virtual reality interventions for personal development: A meta-analysis of hardware and software. Hum. Comput. Interaction 34, 205–239. doi: 10.1080/07370024.2018.1469408

Crossref Full Text | Google Scholar

Huang, Y. K., Mitchell, U. A., Conroy, L. M., and Jones, R. M. (2021). Community daytime noise pollution and socioeconomic differences in Chicago, IL. PLoS One 16:e0254762. doi: 10.1371/journal.pone.0254762

PubMed Abstract | Crossref Full Text | Google Scholar

Jespersen, A. E., Lumbye, A., Schandorff, J., Damgaard, V., Glenthøj, L. B., Nordentoft, M., et al. (2024). Cognition assessment in virtual reality (CAVIR): Associations with neuropsychological performance and activities of daily living in patients with mood or psychosis spectrum disorders. J. Affect. Disord. 369, 1053–1063. doi: 10.1016/j.jad.2024.10.095

PubMed Abstract | Crossref Full Text | Google Scholar

Jeste, D. V., Depp, C. A., and Palmer, B. W. (2006). Magnitude of impairment in decisional capacity in people with schizophrenia compared to normal subjects: An overview. Schizophr. Bull. 32, 121–128. doi: 10.1093/schbul/sbj001

PubMed Abstract | Crossref Full Text | Google Scholar

Ju, Y., Kang, S., Kim, J., Ryu, J. K., and Jeong, E. H. (2024). Clinical Utility of virtual kitchen errand task for children (VKET-C) as a functional cognition evaluation for children with developmental disabilities. Children (Basel) 11:1291. doi: 10.3390/children11111291

PubMed Abstract | Crossref Full Text | Google Scholar

Kählke, F., Buntrock, C., Smit, F., and Ebert, D. D. (2022). Systematic review of economic evaluations for internet- and mobile-based interventions for mental health problems. npj Digital Med. 5:175. doi: 10.1038/s41746-022-00702-w

PubMed Abstract | Crossref Full Text | Google Scholar

Keefe, R. S. E., Davis, V. G., Atkins, A. S., Vaughan, A., Patterson, T., Narasimhan, M., et al. (2016). Validation of a computerized test of functional capacity. Schizophr. Res. 175, 90–96. doi: 10.1016/j.schres.2016.03.038

PubMed Abstract | Crossref Full Text | Google Scholar

Keefe, R. S., Goldberg, T. E., Harvey, P. D., Gold, J. M., Poe, M. P., and Coughenour, L. (2004). The Brief Assessment of Cognition in Schizophrenia: Reliability, sensitivity, and comparison with a standard neurocognitive battery. Schizophr. Res. 68, 283–297. doi: 10.1016/j.schres.2003.09.011

PubMed Abstract | Crossref Full Text | Google Scholar

Keshavan, M. S., and Eack, S. M. (2023). Cognitive enhancement interventions are effective for schizophrenia: Why not provide them early? World Psychiatry 22, 326–327. doi: 10.1002/wps.21091

PubMed Abstract | Crossref Full Text | Google Scholar

Khokhlov, I., Reznik, L., and Ajmera, S. (2020). Sensors in mobile devices knowledge base. IEEE Sensors Lett. 4, 1–4. doi: 10.1109/LSENS.2020.2975161

Crossref Full Text | Google Scholar

Kirkham, R., Kooijman, L., Albertella, L., Myles, D., Yücel, M., and Rotaru, K. (2024). Immersive virtual reality-based methods for assessing executive functioning: Systematic review. JMIR Serious Games 12:e50282. doi: 10.2196/50282

PubMed Abstract | Crossref Full Text | Google Scholar

Kourtesis, P. (2024). A comprehensive review of multimodal XR applications, risks, and ethical challenges in the metaverse. Multimodal Technol. Interact. 8:98. doi: 10.3390/mti8110098

Crossref Full Text | Google Scholar

Kourtis, L. C., Regele, O. B., Wright, J. M., and Jones, G. B. (2019). Digital biomarkers for Alzheimer’s disease: The mobile/wearable devices opportunity. NPJ Digit. Med. 2:9. doi: 10.1038/s41746-019-0084-2

PubMed Abstract | Crossref Full Text | Google Scholar

Kravariti, E., Morris, R. G., Rabe-Hesketh, S., Murray, R. M., and Frangou, S. (2007). Comparative profile analysis of cognitive function in recent-onset and chronic patients with adolescent-onset schizophrenia. Schizophr. Res. 94, 240–244. doi: 10.1016/j.schres.2007.05.014

PubMed Abstract | Crossref Full Text | Google Scholar

Krohn, S., Tromp, J., Quinque, E. M., Belger, J., Klotzsche, F., Rekers, S., et al. (2020). Multidimensional evaluation of virtual reality paradigms in clinical neuropsychology: Application of the VR-check framework. J. Med. Internet Res. 22:e16724. doi: 10.2196/16724

PubMed Abstract | Crossref Full Text | Google Scholar

Kunz, T., Landesvatter, C., and Gummer, T. (2020). Informed consent for paradata use in web surveys. Int. J. Market Res. 62, 396–408. doi: 10.1177/1470785320931669

Crossref Full Text | Google Scholar

Lakhtakia, T., Bondre, A., Chand, P. K., Chaturvedi, N., Choudhary, S., Currey, D., et al. (2022). Smartphone digital phenotyping, surveys, and cognitive assessments for global mental health: Initial data and clinical correlations from an international first episode psychosis study. Digit. Health 8:20552076221133758. doi: 10.1177/20552076221133758

PubMed Abstract | Crossref Full Text | Google Scholar

Lane, E., Joshi, D., Guimond, S., Moore, R., Ahmed, A. O., Ajilore, O., et al. (2024). Exploring current smartphone-based cognitive assessments in schizophrenia and bipolar disorder. Schizophr. Res. Cogn. 37:100309. doi: 10.1016/j.scog.2024.100309

PubMed Abstract | Crossref Full Text | Google Scholar

Lee, S. A., Kim, J. Y., and Park, J. H. (2024). Concurrent validity of virtual reality-based assessment of executive function: A systematic review and meta-analysis. J. Intell. 12:108. doi: 10.3390/jintelligence12110108

PubMed Abstract | Crossref Full Text | Google Scholar

Levaux, M.-N., Potvin, S., Sepehry, A. A., Sablier, J., Mendrek, A., and Stip, E. (2007). Computerized assessment of cognition in schizophrenia: Promises and pitfalls of CANTAB. Eur. Psychiatry 22, 104–115. doi: 10.1016/j.eurpsy.2006.11.004

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, G., Henson, P., Keshavan, M., Pekka-Onnela, J., and Torous, J. (2019). Assessing the potential of longitudinal smartphone based cognitive assessment in schizophrenia: A naturalistic pilot study. Schizophr. Res. Cogn. 17:100144. doi: 10.1016/j.scog.2019.100144

PubMed Abstract | Crossref Full Text | Google Scholar

Madary, M., and Metzinger, T. K. (2016). Real virtuality: A code of ethical conduct. recommendations for good scientific practice and the consumers of VR-technology. Front. Robot. AI 3:3. doi: 10.3389/frobt.2016.00003

Crossref Full Text | Google Scholar

Mancuso, V., Sarcinella, E. D., Bruni, F., Arlati, S., Di Santo, S. G., Cavallo, M., et al. (2024). Systematic review of memory assessment in virtual reality: Evaluating convergent and divergent validity with traditional neuropsychological measures. Front. Hum. Neurosci. 18:1380575. doi: 10.3389/fnhum.2024.1380575

PubMed Abstract | Crossref Full Text | Google Scholar

Martinez-Martin, N. (2024). A broader approach to ethical challenges in digital mental health. World Psychiatry 23, 394–395. doi: 10.1002/wps.21237

PubMed Abstract | Crossref Full Text | Google Scholar

Martinez-Martin, N., Insel, T. R., Dagum, P., Greely, H. T., and Cho, M. K. (2018). Data mining for health: Staking out the ethical territory of digital phenotyping. NPJ Digit. Med. 1:68. doi: 10.1038/s41746-018-0075-8

PubMed Abstract | Crossref Full Text | Google Scholar

McClain, C. A., Couper, M. P., Hupp, A. L., Keusch, F., Peterson, G., Piskorowski, A. D., et al. (2019). A typology of web survey paradata for assessing total survey error. Soc. Sci. Comput. Rev. 37, 196–213. doi: 10.1177/0894439318759670

Crossref Full Text | Google Scholar

Melillo, A., Caporusso, E., Giordano, G. M., Giuliani, L., Pezzella, P., Perrottelli, A., et al. (2023). Correlations between negative symptoms and cognitive deficits in individuals at first psychotic episode or at high risk of psychosis: A systematic review. J. Clin. Med. 12:7095. doi: 10.3390/jcm12227095

PubMed Abstract | Crossref Full Text | Google Scholar

Melillo, A., Chirico, A., De Pietro, G., Gallo, L., Caggianese, G., Barone, D., et al. (2022a). Virtual reality rehabilitation systems for cancer survivors: A narrative review of the literature. Cancers (Basel) 14:3163. doi: 10.3390/cancers14133163

PubMed Abstract | Crossref Full Text | Google Scholar

Melillo, A., Maiorano, P., Rachedi, S., Caggianese, G., Gragnano, E., Gallo, L., et al. (2022b). Labor analgesia: A systematic review and meta-analysis of non-pharmacological complementary and alternative approaches to pain during first stage of labor. Crit. Rev. Eukaryot. Gene Expr. 32, 61–89. doi: 10.1615/CritRevEukaryotGeneExpr.2021039986

PubMed Abstract | Crossref Full Text | Google Scholar

Melillo, A., Perrottelli, A., Caporusso, E., Coltorti, A., Giordano, G. M., Giuliani, L., et al. (2024a). Research evidence on the management of the cognitive impairment component of the post-COVID condition: A qualitative systematic review. Eur. Psychiatry 67:e60. doi: 10.1192/j.eurpsy.2024.1770

PubMed Abstract | Crossref Full Text | Google Scholar

Melillo, A., Rachedi, S., Caggianese, G., Gallo, L., Maiorano, P., Gimigliano, F., et al. (2024b). Synchronization of a virtual reality scenario to uterine contractions for labor pain management: Development study and randomized controlled trial. Games Health J. 13, 389–396. doi: 10.1089/g4h.2023.0202

PubMed Abstract | Crossref Full Text | Google Scholar

Mesholam-Gately, R. I., Giuliano, A. J., Goff, K. P., Faraone, S. V., and Seidman, L. J. (2009). Neurocognition in first-episode schizophrenia: A meta-analytic review. Neuropsychology 23, 315–336. doi: 10.1037/a0014708

PubMed Abstract | Crossref Full Text | Google Scholar

Mesholam-Gately, R. I., Johnston, D., and Keshavan, M. S. (2023). What’s in the name “schizophrenia”? A clinical, research and lived experience perspective. World Psychiatry 22, 156–157. doi: 10.1002/wps.21033

PubMed Abstract | Crossref Full Text | Google Scholar

Miskowiak, K. W., Jespersen, A. E., Kessing, L. V., Aggestrup, A. S., Glenthøj, L. B., Nordentoft, M., et al. (2021). Cognition Assessment in Virtual Reality: Validity and feasibility of a novel virtual reality test for real-life cognitive functions in mood disorders and psychosis spectrum disorders. J. Psychiatr. Res. 145, 182–189. doi: 10.1016/j.jpsychires.2021.12.002

PubMed Abstract | Crossref Full Text | Google Scholar

Moore, R. C., Parrish, E. M., Van Patten, R., Paolillo, E., Filip, T. F., Bomyea, J., et al. (2022). Initial psychometric properties of 7 NeuroUX remote ecological momentary cognitive tests among people with bipolar disorder: Validation study. J. Med. Internet Res. 24:e36665. doi: 10.2196/36665

PubMed Abstract | Crossref Full Text | Google Scholar

Moore, R. C., Swendsen, J., and Depp, C. A. (2017). Applications for self-administered mobile cognitive assessments in clinical research: A systematic review. Int. J. Methods Psychiatr. Res. 26:e1562. doi: 10.1002/mpr.1562

PubMed Abstract | Crossref Full Text | Google Scholar

Negu?, A., Matu, S.-A., Sava, F. A., and David, D. (2016). Virtual reality measures in neuropsychological assessment: A meta-analytic review. Clin. Neuropsychol. 30, 165–184. doi: 10.1080/13854046.2016.1144793

PubMed Abstract | Crossref Full Text | Google Scholar

Neisser, U., and Hyman, I. (2000). Memory Observed: Remembering in Natural Contexts. New York, NY: Macmillan.

Google Scholar

Nguyen, T. M., Leow, A. D., and Ajilore, O. (2023). A review on smartphone keystroke dynamics as a digital biomarker for understanding neurocognitive functioning. Brain Sci. 13:959. doi: 10.3390/brainsci13060959

Crossref Full Text | Google Scholar

Nica, R. (2024). Improving mechanisms of involvement of people with lived experience in decision-making processes. World Psychiatry 23, 398–399. doi: 10.1002/wps.21239

PubMed Abstract | Crossref Full Text | Google Scholar

Nuechterlein, K. H., Green, M. F., and Kern, R. S. (2023). The MATRICS consensus cognitive battery: An update. Curr. Top. Behav. Neurosci. 63, 1–18. doi: 10.1007/7854_2022_395

PubMed Abstract | Crossref Full Text | Google Scholar

Palumbo, D., Caporusso, E., Piegari, G., Mencacci, C., Torriero, S., Giuliani, L., et al. (2022). Social cognition individualized activities lab for social cognition training and narrative enhancement in patients with schizophrenia: A randomized controlled study to assess efficacy and generalization to real-life functioning (Prot. n°: NCT05130853). Front. Psychiatry 13:833550. doi: 10.3389/fpsyt.2022.833550

PubMed Abstract | Crossref Full Text | Google Scholar

Park, J. H. (2022). Can the virtual reality-based spatial memory test better discriminate mild cognitive impairment than neuropsychological assessment? Int. J. Environ. Res. Public Health 19:9950. doi: 10.3390/ijerph19169950

PubMed Abstract | Crossref Full Text | Google Scholar

Parrish, E. M., Kamarsu, S., Harvey, P. D., Pinkham, A., Depp, C. A., and Moore, R. C. (2021). Remote ecological momentary testing of learning and memory in adults with serious mental illness. Schizophr. Bull. 47, 740–750. doi: 10.1093/schbul/sbaa172

PubMed Abstract | Crossref Full Text | Google Scholar

Parsons, T. D. (2015). Virtual reality for enhanced ecological validity and experimental control in the clinical, affective and social neurosciences. Front. Hum. Neurosci. 9:660. doi: 10.3389/fnhum.2015.00660

PubMed Abstract | Crossref Full Text | Google Scholar

Parsons, T. D. (2021). Ethical challenges of using virtual environments in the assessment and treatment of psychopathological disorders. J. Clin. Med. 10:378. doi: 10.3390/jcm10030378

PubMed Abstract | Crossref Full Text | Google Scholar

Perna, G., Caldirola, D., Schatzberg, A. F., and Nemeroff, C. B. (2024). Advancements, challenges and future horizons in personalized psychiatry. World Psychiatry 23, 460–461. doi: 10.1002/wps.21257

PubMed Abstract | Crossref Full Text | Google Scholar

Perrottelli, A., Giordano, G. M., Brando, F., Giuliani, L., Pezzella, P., Mucci, A., et al. (2022). Unveiling the Associations between EEG indices and cognitive deficits in schizophrenia-spectrum disorders: A systematic review. Diagnostics (Basel) 12:2193. doi: 10.3390/diagnostics12092193

PubMed Abstract | Crossref Full Text | Google Scholar

Perrottelli, A., Marzocchi, F. F., Caporusso, E., Giordano, G. M., Giuliani, L., Melillo, A., et al. (2024). Advances in the understanding of the pathophysiology of schizophrenia and bipolar disorder through induced pluripotent stem cell models. J. Psychiatry Neurosci. 49, E109–E125. doi: 10.1503/jpn.230112

PubMed Abstract | Crossref Full Text | Google Scholar

Petersen, J. Z., Porter, R. J., and Miskowiak, K. W. (2019). Clinical characteristics associated with the discrepancy between subjective and objective cognitive impairment in depression. J. Affect. Disord. 246, 763–774. doi: 10.1016/j.jad.2018.12.105

PubMed Abstract | Crossref Full Text | Google Scholar

Pezzella, P., Caporusso, E., Mucci, A., Bucci, P., Giordano, G. M., Amore, M., et al. (2024). Interview versus performance assessment of cognition as predictors of real-world outcomes in a large-scale cross-sectional study in schizophrenia. Schizophr. Bull. Open 5:sgae020. doi: 10.1093/schizbullopen/sgae020

PubMed Abstract | Crossref Full Text | Google Scholar

Rasmussen, L. J. H., Caspi, A., Ambler, A., Broadbent, J. M., Cohen, H. J., d’Arbeloff, T., et al. (2019). Association of neurocognitive and physical function with gait speed in midlife. JAMA Netw. Open 2:e1913123. doi: 10.1001/jamanetworkopen.2019.13123

PubMed Abstract | Crossref Full Text | Google Scholar

Reiter, T., and Schoedel, R. (2024). Never miss a beep: Using mobile sensing to investigate (non-)compliance in experience sampling studies. Behav. Res. Methods 56, 4038–4060. doi: 10.3758/s13428-023-02252-9

PubMed Abstract | Crossref Full Text | Google Scholar

Riva, G., Malighetti, C., and Serino, S. (2021). Virtual reality in the treatment of eating disorders. Clin. Psychol. Psychother. 28, 477–488. doi: 10.1002/cpp.2622

PubMed Abstract | Crossref Full Text | Google Scholar

Robbins, T. W., James, M., Owen, A. M., Sahakian, B. J., McInnes, L., and Rabbitt, P. (2010). Cambridge neuropsychological test automated battery (CANTAB): A factor analytic study of a large sample of normal elderly volunteers. Dementia 5, 266–281. doi: 10.1159/000106735

PubMed Abstract | Crossref Full Text | Google Scholar

Rosenberg, R. S., Baughman, S. L., and Bailenson, J. N. (2013). Virtual superheroes: Using superpowers in virtual reality to encourage prosocial behavior. PLoS One 8:e55003. doi: 10.1371/journal.pone.0055003

PubMed Abstract | Crossref Full Text | Google Scholar

Rousseaux, F., Faymonville, M.-E., Nyssen, A.-S., Dardenne, N., Ledoux, D., Massion, P. B., et al. (2020). Can hypnosis and virtual reality reduce anxiety, pain and fatigue among patients who undergo cardiac surgery: A randomised controlled trial. Trials 21:330. doi: 10.1186/s13063-020-4222-6

PubMed Abstract | Crossref Full Text | Google Scholar

Rucci, P., Caporusso, E., Sanmarchi, F., Giordano, G. M., Mucci, A., Giuliani, L., et al. (2023). The structure stability of negative symptoms: Longitudinal network analysis of the brief negative symptom scale in people with schizophrenia. BJPsych Open 9:e168. doi: 10.1192/bjo.2023.541

PubMed Abstract | Crossref Full Text | Google Scholar

Rus-Calafell, M., Garety, P., Sason, E., Craig, T. J. K., and Valmaggia, L. R. (2018). Virtual reality in the assessment and treatment of psychosis: A systematic review of its utility, acceptability and effectiveness. Psychol. Med. 48, 362–391. doi: 10.1017/s0033291717001945

PubMed Abstract | Crossref Full Text | Google Scholar

Sahakian, B. J., and Owen, A. (1992). Computerized assessment in neuropsychiatry using CANTAB: Discussion paper. J. R. Soc. Med. 85, 399–402.

Google Scholar

Sahoo, S., and Grover, S. (2022). Technology-based neurocognitive assessment of the elderly: A mini review: Оценка нейрокогнитивных функций на основе компьютерных технологий у пожилых людей: краткий обзор. Consort Psychiatr. 3, 37–44. doi: 10.17816/cp155

PubMed Abstract | Crossref Full Text | Google Scholar

Sakdalan, J., and Mitchell, B. (2023). Development of an evidence-based violence rehabilitation program for offenders with intellectual disability. Int. J. Offender Ther. Comp. Criminol. 69, 572–593. doi: 10.1177/0306624X231176005

PubMed Abstract | Crossref Full Text | Google Scholar

Sampogna, G., Di Vincenzo, M., Giuliani, L., Menculini, G., Mancuso, E., Arsenio, E., et al. (2023). A systematic review on the effectiveness of antipsychotic drugs on the quality of life of patients with schizophrenia. Brain Sci. 13:1577. doi: 10.3390/brainsci13111577

PubMed Abstract | Crossref Full Text | Google Scholar

Sampogna, G., Toni, C., Catapano, P., Rocca, B. D., Di Vincenzo, M., Luciano, M., et al. (2024). New trends in personalized treatment of depression. Curr. Opin. Psychiatry 37, 3–8. doi: 10.1097/yco.0000000000000903

PubMed Abstract | Crossref Full Text | Google Scholar

Sansone, N., Tyano, S., Melillo, A., Schouler-Ocak, M., and Galderisi, S. (2024). Comparing the world psychiatric association and european psychiatric association codes of ethics: Discrepancies and shared grounds. Eur. Psychiatry 67:e38. doi: 10.1192/j.eurpsy.2024.1748

PubMed Abstract | Crossref Full Text | Google Scholar

Schäfer, S. K., Thomas, L. M., Lindner, S., and Lieb, K. (2023). World Health Organization’s low-intensity psychosocial interventions: A systematic review and meta-analysis of the effects of Problem Management Plus and Step-by-Step. World Psychiatry 22, 449–462. doi: 10.1002/wps.21129

PubMed Abstract | Crossref Full Text | Google Scholar

Shvetz, C., Gu, F., Drodge, J., Torous, J., and Guimond, S. (2021). Validation of an ecological momentary assessment to measure processing speed and executive function in schizophrenia. NPJ Schizophr. 7:64. doi: 10.1038/s41537-021-00194-9

PubMed Abstract | Crossref Full Text | Google Scholar

Singh, S., Strong, R., Xu, I., Fonseca, L. M., Hawks, Z., Grinspoon, E., et al. (2023). Ecological momentary assessment of cognition in clinical and community samples: Reliability and validity study. J. Med. Internet Res. 25:e45028. doi: 10.2196/45028

PubMed Abstract | Crossref Full Text | Google Scholar

Smyth, J. M., and Ebner-Priemer, U. (2025). Dispelling “pleasing myths” about the integration of ecological momentary assessment and intervention into clinical research and practice. World Psychiatry 24, 2–3. doi: 10.1002/wps.21262

PubMed Abstract | Crossref Full Text | Google Scholar

Spieker, E. A., Astur, R. S., West, J. T., Griego, J. A., and Rowland, L. M. (2012). Spatial memory deficits in a virtual reality eight-arm radial maze in schizophrenia. Schizophr. Res. 135, 84–89. doi: 10.1016/j.schres.2011.11.014

PubMed Abstract | Crossref Full Text | Google Scholar

Staal, W. G., Hijman, R., Hulshoff Pol, H. E., and Kahn, R. S. (2000). Neuropsychological dysfunctions in siblings discordant for schizophrenia. Psychiatry Res. 95, 227–235. doi: 10.1016/s0165-1781(00)00172-4

PubMed Abstract | Crossref Full Text | Google Scholar

Starzer, M., Hansen, H. G., Hjorthøj, C., Albert, N., Nordentoft, M., and Madsen, T. (2023). 20-year trajectories of positive and negative symptoms after the first psychotic episode in patients with schizophrenia spectrum disorder: Results from the OPUS study. World Psychiatry 22, 424–432. doi: 10.1002/wps.21121

PubMed Abstract | Crossref Full Text | Google Scholar

Steindl, E. (2023). Safeguarding privacy and efficacy in e-mental health: Policy options in the EU and Australia. Int. Data Privacy Law 13, 207–224. doi: 10.1093/idpl/ipad009

Crossref Full Text | Google Scholar

Stone, A. A., Schneider, S., and Smyth, J. M. (2023). Evaluation of pressing issues in ecological momentary assessment. Annu. Rev. Clin. Psychol. 19, 107–131. doi: 10.1146/annurev-clinpsy-080921-083128

PubMed Abstract | Crossref Full Text | Google Scholar

Sumiyoshi, T., Watanabe, K., Noto, S., Sakamoto, S., Moriguchi, Y., Tan, K. H. X., et al. (2019). Relationship of cognitive impairment with depressive symptoms and psychosocial function in patients with major depressive disorder: Cross-sectional analysis of baseline data from PERFORM-J. J. Affect. Disord. 258, 172–178. doi: 10.1016/j.jad.2019.07.064

PubMed Abstract | Crossref Full Text | Google Scholar

Sweere, D. J. J., Pel, J. J. M., Kooiker, M. J. G., van Dijk, J. P., van Gemert, E., Hurks, P. P. M., et al. (2022). Clinical utility of eye tracking in assessing distractibility in children with neurological disorders or ADHD: A cross-sectional study. Brain Sci. 12:1369. doi: 10.3390/brainsci12101369

PubMed Abstract | Crossref Full Text | Google Scholar

Tamminga, C. A. (2023). Discovering informative biomarkers in psychiatry. World Psychiatry 22, 270–271. doi: 10.1002/wps.21084

PubMed Abstract | Crossref Full Text | Google Scholar

Terhorst, Y., Knauer, J., Philippi, P., and Baumeister, H. (2024). The relation between passively collected GPS mobility metrics and depressive symptoms: Systematic review and meta-analysis. J. Med. Internet Res. 26:e51875. doi: 10.2196/51875

PubMed Abstract | Crossref Full Text | Google Scholar

Torous, J., Bucci, S., Bell, I. H., Kessing, L. V., Faurholt-Jepsen, M., Whelan, P., et al. (2021). The growing field of digital psychiatry: Current evidence and the future of apps, social media, chatbots, and virtual reality. World Psychiatry 20, 318–335. doi: 10.1002/wps.20883

PubMed Abstract | Crossref Full Text | Google Scholar

Tyano, S. (2024). Ethics from the lens of the social dimension of psychiatry. World Psychiatry 23, 396–397. doi: 10.1002/wps.21238

PubMed Abstract | Crossref Full Text | Google Scholar

Ventura, J., Reise, S. P., Keefe, R. S., Baade, L. E., Gold, J. M., Green, M. F., et al. (2010). The cognitive assessment interview (CAI): Development and validation of an empirically derived, brief interview-based measure of cognition. Schizophr. Res. 121, 24–31. doi: 10.1016/j.schres.2010.04.016

PubMed Abstract | Crossref Full Text | Google Scholar

Ventura, J., Welikson, T., Ered, A., Subotnik, K. L., Keefe, R. S. E., Hellemann, G. S., et al. (2020). Virtual reality assessment of functional capacity in the early course of schizophrenia: Associations with cognitive performance and daily functioning. Early Interv. Psychiatry 14, 106–114. doi: 10.1111/eip.12831

PubMed Abstract | Crossref Full Text | Google Scholar

Vita, A., Gaebel, W., Mucci, A., Sachs, G., Barlati, S., Giordano, G. M., et al. (2022a). European psychiatric association guidance on treatment of cognitive impairment in schizophrenia. Eur. Psychiatry 65:e57. doi: 10.1192/j.eurpsy.2022.2315

PubMed Abstract | Crossref Full Text | Google Scholar

Vita, A., Gaebel, W., Mucci, A., Sachs, G., Erfurth, A., Barlati, S., et al. (2022b). European psychiatric association guidance on assessment of cognitive impairment in schizophrenia. Eur. Psychiatry 65:e58. doi: 10.1192/j.eurpsy.2022.2316

PubMed Abstract | Crossref Full Text | Google Scholar

Vita, A., Mussoni, C., Deste, G., Ferlenghi, G., Turrina, C., and Valsecchi, P. (2018). Psychopharmacological treatment of cognitive deficits in Schizophrenia and mood disorders. Official J. Italian Soc. Psychopathol. 24, 62–72.

Google Scholar

Voineskos, A. (2023). The time is now to start evaluating biomarkers in the clinic. World Psychiatry 22, 267–268. doi: 10.1002/wps.21082

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, S. B., Wang, Y. Y., Ungvari, G. S., Ng, C. H., Wu, R. R., Wang, J., et al. (2017). The macarthur competence assessment tools for assessing decision-making capacity in schizophrenia: A meta-analysis. Schizophr. Res. 183, 56–63. doi: 10.1016/j.schres.2016.11.020

PubMed Abstract | Crossref Full Text | Google Scholar

Weizenbaum, E., Torous, J., and Fulford, D. (2020). Cognition in context: Understanding the everyday predictors of cognitive performance in a new era of measurement. JMIR Mhealth Uhealth 8:e14328. doi: 10.2196/14328

PubMed Abstract | Crossref Full Text | Google Scholar

Wiebe, A., Kannen, K., Selaskowski, B., Mehren, A., Thöne, A. K., Pramme, L., et al. (2022). Virtual reality in the diagnostic and therapy for mental disorders: A systematic review. Clin. Psychol. Rev. 98:102213. doi: 10.1016/j.cpr.2022.102213

PubMed Abstract | Crossref Full Text | Google Scholar

Wiener, J. M., Carroll, D., Moeller, S., Bibi, I., Ivanova, D., Allen, P., et al. (2020). A novel virtual-reality-based route-learning test suite: Assessing the effects of cognitive aging on navigation. Behav. Res. Methods 52, 630–640. doi: 10.3758/s13428-019-01264-8

PubMed Abstract | Crossref Full Text | Google Scholar

Wilkins, L. K., Girard, T. A., King, J., King, M. J., Herdman, K. A., Christensen, B. K., et al. (2013). Spatial-memory deficit in schizophrenia spectrum disorders under viewpoint-independent demands in the virtual courtyard task. J. Clin. Exp. Neuropsychol. 35, 1082–1093. doi: 10.1080/13803395.2013.857389

PubMed Abstract | Crossref Full Text | Google Scholar

Wittkopf, P. G., Lloyd, D. M., Coe, O., Yacoobali, S., and Billington, J. (2020). The effect of interactive virtual reality on pain perception: A systematic review of clinical studies. Disabil. Rehabil. 42, 3722–3733. doi: 10.1080/09638288.2019.1610803

PubMed Abstract | Crossref Full Text | Google Scholar

Wrzus, C., and Neubauer, A. B. (2023). Ecological momentary assessment: A meta-analysis on designs, samples, and compliance across research fields. Assessment 30, 825–846. doi: 10.1177/10731911211067538

PubMed Abstract | Crossref Full Text | Google Scholar

Wu, Y. H., Xu, J., Stangl, E., Pentony, S., Vyas, D., Chipara, O., et al. (2021). Why ecological momentary assessment surveys go incomplete: When it happens and how it impacts data. J. Am. Acad. Audiol. 32, 16–26. doi: 10.1055/s-0040-1719135

PubMed Abstract | Crossref Full Text | Google Scholar

Yaroslavsky, I., Allard, E. S., and Sanchez-Lopez, A. (2019). Can’t look Away: Attention control deficits predict rumination, depression symptoms and depressive affect in daily life. J. Affect. Disord. 245, 1061–1069. doi: 10.1016/j.jad.2018.11.036

PubMed Abstract | Crossref Full Text | Google Scholar

Yatham, L. N., Torres, I. J., Malhi, G. S., Frangou, S., Glahn, D. C., Bearden, C. E., et al. (2010). The international society for bipolar disorders-battery for assessment of neurocognition (ISBD-BANC). Bipolar Disord. 12, 351–363. doi: 10.1111/j.1399-5618.2010.00830.x

PubMed Abstract | Crossref Full Text | Google Scholar

Zamora, W., Calafate, C. T., Cano, J. C., and Manzoni, P. (2017). Accurate ambient noise assessment using smartphones. Sensors (Basel) 17:917. doi: 10.3390/s17040917

PubMed Abstract | Crossref Full Text | Google Scholar

Zimmerman, M. (2024). The value and limitations of self-administered questionnaires in clinical practice and epidemiological studies. World Psychiatry 23, 210–212. doi: 10.1002/wps.21191

PubMed Abstract | Crossref Full Text | Google Scholar

Zulueta, J., Piscitello, A., Rasic, M., Easter, R., Babu, P., Langenecker, S. A., et al. (2018). Predicting mood disturbance severity with mobile phone keystroke metadata: A biaffect digital phenotyping study. J. Med. Internet Res. 20:e241. doi: 10.2196/jmir.9775

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: cognitive assessment, ecological momentary assessments, virtual reality, digital phenotyping, severe mental illness, schizophrenia, wearable devices, digital medicine

Citation: Caporusso E, Melillo A, Perrottelli A, Giuliani L, Marzocchi FF, Pezzella P and Giordano GM (2025) Current limitations in technology-based cognitive assessment for severe mental illnesses: a focus on feasibility, reliability, and ecological validity. Front. Behav. Neurosci. 19:1543005. doi: 10.3389/fnbeh.2025.1543005

Received: 10 December 2024; Accepted: 24 March 2025;
Published: 07 April 2025.

Edited by:

Simona Cabib, Santa Lucia Foundation (IRCCS), Italy

Reviewed by:

Laura Piccardi, Sapienza University of Rome, Italy
Valentina Mancuso, eCampus University, Italy
Martino Schettino, Sapienza University of Rome, Italy

Copyright © 2025 Caporusso, Melillo, Perrottelli, Giuliani, Marzocchi, Pezzella and Giordano. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Andrea Perrottelli, YW5kcmVhLnBlcnJvdHRlbGxpQHVuaWNhbXBhbmlhLml0

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.