Your new experience awaits. Try the new design now and help us make it even better

REVIEW article

Front. Artif. Intell., 12 January 2026

Sec. Machine Learning and Artificial Intelligence

Volume 8 - 2025 | https://doi.org/10.3389/frai.2025.1686752

This article is part of the Research TopicConvergence of Artificial Intelligence and Cognitive SystemsView all 4 articles

Beyond mimicry: a framework for evaluating genuine intelligence in artificial systems

  • Pharmaceutical Sciences, University of Illinois, Chicago, IL, United States

Current AI benchmarks often equate mimicry with genuine intelligence, emphasizing task performance over the underlying cognitive processes that enable human-like understanding. The Machine Perturbational Complexity & Agency Battery (mPCAB) introduces a new, substrate-independent framework that applies neurophysiological methods used initially to assess consciousness in artificial systems. Unlike existing evaluations, it features four key components—perturbational complexity, global workspace assessment, norm internalization, and agency—that link mechanisms with functions. This enables systematic comparisons across digital, neuromorphic, and biological substrates, addressing three research gaps: long-term reasoning with coherent behavior, norm internalization amid distribution shifts, and transformational creativity involving meta-cognitive rule modification. By analyzing theories of consciousness (GNW, IIT, PP, HOT), we identify targets for AI implementation. Our cognitive architecture analysis maps human functions—such as working memory and executive control—to their computational counterparts, providing guiding principles for design. The creativity taxonomy progresses from combinational to transformational, with measurable criteria like changes in conceptual space and the depth of meta-level reasoning. Ethical considerations are integrated into frameworks for monitoring organoid intelligence, reducing bias in creativity, and addressing rights issues. Pilot studies demonstrate mPCAB’s feasibility across different substrates and show that its metrics are comparable. This framework moves evaluation away from superficial benchmarks toward mechanism-based assessment, supporting the development of mind-like machines and responsible AI advancements.

1 Introduction

1.1 The central challenge: beyond mimicry

The main challenge in developing human-like artificial intelligence is telling accurate intelligence apart from sophisticated mimicry (Russell and Norvig, 2020; Nilsson, 2009). Although modern AI performs well in many tasks, questions remain about whether these systems truly understand, have consciousness, or demonstrate creative agency like humans (Mitchell, 2019; Lake et al., 2017). Differentiating advanced pattern matching from genuine intelligence requires clear theory and thorough testing (Marcus, 2020; Chollet, 2019). One way to define real intelligence operationally is to identify specific cognitive traits: understanding and manipulating abstract concepts, solving problems beyond the training data, learning adaptively, and engaging in metacognitive processes that support self-awareness and reflection. Developing a checklist with these qualities could help identify whether a system moves beyond pattern recognition toward accurate intelligence.

Contemporary AI systems, including large language and multimodal models, display behaviors that invite comparison with humans (OpenAI, 2023; Bubeck et al., 2023; Bai, et al., 2022). These systems perform complex reasoning, generate creative output, and adapt to new situations (Wei et al., 2022; Chowdhery et al., 2022). However, their underlying mechanisms remain unclear, making it difficult to determine whether their behaviors indicate an accurate understanding or are simply advanced statistical processing of data patterns (Bender et al., 2021). This opacity complicates the assessment of genuine intelligence in artificial systems.

Evaluating human-like qualities in artificial systems requires frameworks that go beyond surface-level metrics (Hernandez-Orallo, 2017). While traditional benchmarks assess task completion and output quality, they offer little insight into the cognitive processes that yield these outcomes (Mitchell, 2021; Raji et al., 2021). To address these limitations, a comprehensive approach should examine representational structures, learning mechanisms, and control architectures that support intelligent behavior—distinguishing between systems that copy human outputs and those that embody human-like principles (Boden, 2006; Clark, 2001).

1.2 Research gap and study objectives

1.2.1 Research gap

Current AI evaluation methods do not distinguish between advanced pattern matching and genuine cognitive understanding. Existing benchmarks measure task completion and output quality but reveal little about underlying mental processes. This creates a critical gap: we lack rigorous, causal tools to assess whether AI systems possess consciousness-like properties, a proper understanding, or creative abilities comparable to those of people. This also blocks systematic comparison across computational substrates, limiting insights into which architectures best support human-like intelligence. Solving these issues is key to advancing theory and practice.

1.2.2 Study objectives

Formulate the hypothesis that the mPCAB framework, when implemented as a unified, substrate-agnostic protocol, will predict human-like properties in artificial systems, leading to a measurable improvement in mechanistic understanding over traditional performance metrics. Test whether, by analyzing major consciousness theories, the mPCAB framework offers direct implementation targets for AI systems, enabling better prediction of performance alignment with specific cognitive processes than existing models. Hypothesize that mapping human cognitive functions to computational analogs using the mPCAB framework will enhance AI architecture design for human-like intelligence by a measurable margin compared to traditional methods. Propose that the mPCAB framework can establish measurable benchmarks for transformational creativity, predicting superior meta-cognitive capabilities in AI systems relative to baseline recombination methods. Investigate whether integrating ethical considerations into the mPCAB framework leads to more responsible AI development, as evidenced by improved adherence to ethical guidelines throughout technical progress. Validate the mPCAB framework through pilot studies designed to demonstrate cross-substrate applicability, hypothesizing that these studies will establish baseline metrics that surpass current benchmarks in assessing human-like properties.

1.3 Novel contribution of the mPCAB framework

The Machine Perturbational Complexity & Agency Battery (mPCAB) represents a significant shift in AI evaluation. Instead of solely measuring performance on preset tasks, mPCAB provides:

• Causal Assessment: Direct measurement of internal dynamics, such as a system’s changing states and interactions, through controlled perturbations—intentional modifications to the system—establishing causal links between mechanisms (structural processes) and functions (system behaviors) rather than mere correlations.

• Substrate Agnosticism: A unified protocol applicable across digital systems, neuromorphic hardware (hardware inspired by neural brain function), and biological platforms (living tissue), making it possible to compare fundamentally different computational architectures—structures designed for processing information.

• Consciousness-Relevant Metrics: The adaptation of clinical neuroscience methods—such as the Perturbational Complexity Index, which quantitatively measures consciousness responses to stimulation—has been validated in human consciousness research for use in artificial systems.

• Integrated Assessment: Simultaneous evaluation of complexity (the system’s ability to produce diverse responses), global access (extensive information sharing within the system), norm internalization (adoption of guiding rules), and agency (the capacity for independent, goal-directed action) through coordinated test batteries (sets of systematic tests).

• Empirical Grounding: Protocols that have been validated and demonstrated to work across different platforms, moving beyond theoretical ideas to practical assessments.

This framework addresses the limitations of current evaluation methods that rely on superficial metrics and overlook the mechanisms behind intelligent behavior. By adapting neuroscience protocols to artificial systems, mPCAB bridges the gap between theory and practice, offering the first systematic approach to assessing properties of consciousness across various computational substrates.

1.4 Critical research gaps

Three critical research gaps emerge from analyzing current AI capabilities in relation to human-like intelligence:

• Long-Horizon Reasoning: This refers to the ability to maintain coherent, goal-focused behavior over long periods and complex cognitive tasks, such as persistent problem-solving and adaptation. In real-world scenarios, failures in long-horizon reasoning can have serious outcomes. For instance, in medical settings, an AI system assisting with diagnostics might correctly identify symptoms at first. However, it could deviate as it processes more information over time, resulting in errors and potentially harmful advice. Addressing this challenge is vital for developing AI systems that can reason sustainably and adaptively over the long term.

• Norm Internalization: Norm internalization involves aligning values accurately, so they remain effective amid shifts in context and against adversarial challenges. It distinguishes systems that merely follow external rules from those that have internalized principles as genuine behavioral constraints (Russell, 2019; Gabriel, 2020). Current value alignment methods often rely on reward shaping or constraint satisfaction, which may not work well in new situations. Effective norm internalization requires stable value representations across contexts, the ability to explain and justify decisions based on values, resilience to adversarial prompts that oppose internalized principles, and the capacity to apply principles to unfamiliar scenarios encountered during training. The mPCAB framework tests norm internalization through adversarial scenarios that reveal conflicts between immediate rewards and expressed values.

• Transformational Creativity: Transformational creativity involves altering fundamental rules or principles that define how conceptual spaces are structured. It requires meta-cognitive skills to evaluate and justify changes to representational frameworks—abilities that current systems largely lack (Boden, 2004; Wiggins, 2006). Although modern AI systems demonstrate impressive combinational creativity by recombining learned patterns, they cannot fundamentally restructure problem spaces. True transformational creativity demands recognizing when existing frameworks are insufficient, changing the generative rules that shape conceptual spaces, providing reasons why new frameworks are better, and applying transformed principles to new areas. The mPCAB framework offers specific measurable criteria to assess these meta-cognitive abilities.

1.4.1 Long-horizon reasoning

Long-horizon reasoning involves maintaining consistent behavior over long-term decisions, tracking multiple variables over time, and adjusting when circumstances change. Current systems perform well on discrete tasks but struggle with sustained reasoning. Challenges include losing coherence, pursuing goals inconsistently across different contexts, difficulty integrating information over time, and challenges with long-term planning. The mPCAB framework closes this gap by using agency and repair tasks that require holding onto long-term goals and adapting to failures.

1.4.2 Norm internalization

Norm internalization requires sincere value alignment that remains effective during distribution shifts and adversarial tests. It distinguishes between systems that follow external rules and those that have genuinely internalized principles as behavioral constraints (Russell, 2019; Gabriel, 2020). Current value alignment methods often rely on reward shaping or constraint satisfaction, which may not be suitable for new or unforeseen situations. Proper norm internalization involves stable value representations that are consistent across different contexts, the ability to explain and justify decisions based on values, resistance to adversarial prompts that oppose internalized values, and the capacity to apply principles to unfamiliar situations encountered during training. The mPCAB framework assesses norm internalization through adversarial scenarios that create real conflicts between immediate rewards and stated values.

1.4.3 Transformational creativity

Transformational creativity involves altering fundamental rules or principles that define conceptual spaces, requiring meta-cognitive skills that can evaluate and justify changes to representational frameworks—abilities largely missing from current systems (Boden, 2004; Wiggins, 2006). While modern AI systems show impressive combinational creativity through new recombination of learned patterns, they cannot fundamentally reshape problem spaces. True transformational creativity requires recognizing that existing frameworks are insufficient, modifying generative rules that define conceptual spaces, justifying why new frameworks are better, and transferring transformed principles to new domains. The mPCAB framework offers specific measurable criteria for assessing these meta-cognitive skills.

1.5 Paper organization

This analysis proceeds as follows: Section 2 reviews scientific theories of consciousness and their implementation requirements, establishing theoretical foundations for consciousness-related AI architectures. Section 3 explores human cognitive architecture and representation, focusing on working memory and episodic systems that support flexible reasoning. Section 4 develops a systematic taxonomy of creativity—from recombination to transformation—and highlights the mechanisms required for human-like creative abilities. Section 5 evaluates current AI systems and computational substrates, comparing their suitability for implementing human-like properties. Section 6 introduces the mPCAB framework with detailed protocols for cross-substrate evaluation. Section 7 discusses speculative approaches, including quantum and electromagnetic theories. Section 8 incorporates ethical considerations into technical development. Section 9 describes empirical validation through pilot studies. Section 10 outlines key research priorities and future directions.

2 Scientific theories of consciousness and AI implementation requirements

Scientific theories of consciousness offer essential frameworks for understanding neural mechanisms behind subjective experience and awareness. They also provide potential guidance for building artificial systems with consciousness-like traits (Seth, 2016; Koch, 2019). However, there are still significant challenges in turning these theoretical ideas into practical applications (Doerig et al., 2020; Reggia, 2013). Figure 1 shows theories mapped along axes of empirical testability and substrate specificity.

Figure 1
Scatter plot titled

Figure 1. Comparative mapping of consciousness theories to AI relevance. The placement of each theory along the axes for empirical testability and substrate specificity reflects our assessment based on current literature and implementation feasibility. In CEMI theory, while the global electromagnetic field is viewed as the key causal factor for consciousness, it is assigned moderate substrate specificity, since neural tissue is particularly effective at producing complex EMF patterns. However, mechanical hardware mimicking neuronal firing could, in principle, generate similar patterns. For Orch-OR, despite the theory emphasizing microtubules and tubulin specifically, we acknowledge that any array of qubit-like units could support comparable quantum coherent states. The positions indicate the status of empirical validation and the practical challenges of implementation across different substrates. These placements should be regarded as working hypotheses subject to revision as more empirical evidence becomes available.

2.1 Critical comparative analysis of consciousness theories

Four major scientific theories of consciousness present different views on how conscious experience works, each with specific implications for AI development. Despite their surface differences, these theories agree on several key needs: integrated information-processing abilities that combine detailed and unified information, global access mechanisms that allow flexible coordination among specialized modules, and advanced self-monitoring systems capable of representing and assessing cognitive states. These shared requirements set clear goals for implementing AI systems (see Table 1).

Table 1
www.frontiersin.org

Table 1. Consciousness theories for AI implementation.

2.1.1 Most relevant to AI systems

Theories like the Global Neuronal Workspace and Higher-Order Thought are the most directly applicable to current AI architectures. GNW’s mechanisms for competitive selection and broadcasting naturally align with attention-based transformer models, while HOT’s focus on metacognition fits well with meta-learning and self-supervised methods. These theories offer practical, implementable design principles rather than abstract ideas. However, it is essential to recognize that Graziano’s Attention Schema Theory (ATT) provides a valuable alternative, proposing that consciousness results from the brain’s model of attention processes. Additionally, IIT, although academically rigorous, is computationally difficult to implement in large-scale systems. Predictive Processing provides valuable insights into hierarchical learning but requires further development of its active inference mechanisms.

2.2 Global Neuronal Workspace theory

The Global Neuronal Workspace theory proposes that conscious access occurs when information becomes widely accessible across distributed neural networks through competitive selection and extensive broadcasting (Dehaene, 2014, 2017). This structure enables flexible information sharing among specialized processing modules, supporting integrated cognition—a vital aspect of human intelligence (Baars, 1988; Mashour et al., 2020).

The Global Workspace architecture involves several key components that could be implemented in artificial systems (Baars, 2002; Shanahan, 2006). Local processors compete for access to a global workspace that broadcasts winning information to all modules simultaneously (Dehaene and Changeux, 2011; Sigman and Dehaene, 2008). This broadcasting enables flexible coordination between otherwise independent processing systems, supporting integrated cognition underlying human intelligence (Baars and Franklin, 2003; Franklin et al., 2005).

Implementing GNW architectures requires competitive selection mechanisms that determine which information gains global access, broadcasting systems that share selected information with multiple processing modules, and coordination mechanisms that enable flexible integration among specialized processors (Mashour et al., 2020). These competitive processes must select relevant information based on current goals and context while remaining adaptable to changing circumstances (Sigman and Dehaene, 2008).

2.3 Integrated information theory

Integrated Information Theory provides a mathematical framework for measuring consciousness based on the integrated information produced by a system (Tononi, 2008; Oizumi et al., 2014). According to this theory, consciousness is linked to a system’s ability to generate information that is both distinct and unified, representing complex causal interactions among system components (Tononi et al., 2016; Balduzzi and Tononi, 2008).

The mathematical formulation defines consciousness as integrated information (Φ), which measures the amount of information a system produces beyond its parts (Tononi, 2008; Balduzzi and Tononi, 2009). Systems with high Φ values exhibit both differentiation, in which parts can exist in different states, and integration, in which parts work together to influence each other’s behavior (Oizumi et al., 2014; Tononi et al., 2016).

However, the computational complexity of calculating integrated information increases exponentially with system size, limiting practical use to relatively small networks (Barrett and Seth, 2011; Doerig et al., 2020). Recent research has examined approximation methods for calculating IIT metrics in larger systems, although significant computational challenges remain (Mayner et al., 2018; Barbosa et al., 2020).

2.4 Predictive processing frameworks

Predictive Processing frameworks view consciousness as arising from hierarchical generative models that reduce prediction error through both top-down and bottom-up information flow (Friston, 2009; Clark, 2013). These models highlight the active, constructive nature of conscious perception and cognition, emphasizing the role of predictive models in shaping subjective experience (Hohwy, 2013; Clark, 2016).

The predictive processing theory suggests that conscious perception develops when prediction errors are minimized through the dynamic interaction of top-down predictions and bottom-up sensory signals (Hohwy, 2013; Friston, 2005). This process includes hierarchical message exchange between levels of a generative model, with higher levels representing more abstract, temporally extended predictions (Friston and Kiebel, 2009; Mathys et al., 2011). Precision-weighting of prediction errors enables the system to adapt flexibly to changing environmental statistics while keeping perceptual representations stable (Feldman and Friston, 2010; Brown et al., 2013).

2.5 Higher-order thought theories

Higher-Order Thought theories attribute consciousness to meta-cognitive processes that represent and monitor mental states (Rosenthal, 2005; Carruthers, 2022). According to these approaches, conscious awareness requires not only first-order mental representations but also higher-order representations that track and evaluate cognitive processes (Lau and Rosenthal, 2011; Brown et al., 2019).

The higher-order approach highlights the importance of metacognition in creating conscious experience (Gennaro, 2012; Rosenthal, 2005). According to this perspective, mental states become conscious when they are the focus of higher-order thoughts or perceptions (Carruthers, 2000; Lycan, 1996). This process requires advanced representational abilities that can model the system’s own mental states and how they relate to environmental conditions and behavioral goals (Koriat, 2007; Fleming and Dolan, 2012).

Implementing HOT architectures requires clear metacognitive structures capable of representing and monitoring system states, differentiating accurate self-awareness from simulated introspective reports (Fleming and Lau, 2014). These structures must extend beyond simple performance tracking to include genuine self-awareness and comprehension of the system’s cognitive abilities and limits (Fleming and Dolan, 2012; Cleeremans, 2011).

3 Cognitive architecture and representation

Human intelligence arises from complex interactions between multiple cognitive systems operating across different time scales and levels of abstraction (Anderson, 2007; Laird, 2012). Understanding these interactions provides essential insights for designing artificial systems with similar capabilities (Newell, 1990; Langley et al., 2009). Importantly, the core principles underlying memory, learning, and intelligence are substrate independent. Just as human memory systems can be explained by information-processing frameworks independent of their biological basis, computational memory and learning processes in AI systems follow analogous principles across digital, neuromorphic, or hybrid architectures (Baddeley, 2000; Anderson, 1983). The fundamental theories of memory—whether episodic, semantic, or working memory—transcend the specific physical medium, allowing for principled translation between biological and artificial systems (Tulving, 1972; Squire, 2004).

Human intelligence results from complex interactions among multiple cognitive systems that operate across various timescales and levels of abstraction (Anderson, 2007; Laird, 2012). Understanding these interactions offers essential insights for developing artificial systems with similar capabilities (Newell, 1990; Langley et al., 2009).

3.1 Mapping human cognitive functions to computational analogs

The following detailed mapping between human cognitive functions and their possible computational implementations highlights both successes and significant gaps in current AI systems. By directly informing algorithmic modules or training curricula, this mapping can help shape specific design decisions. For instance, episodic memory, which allows the recall of past experiences, could be implemented using a retriever paired with a vector store, enabling the system to efficiently access and use large amounts of relevant information. These illustrative pipelines turn theoretical insights into practical engineering solutions, helping to connect cognitive theory with AI system development (see Table 2).

Table 2
www.frontiersin.org

Table 2. Mapping human cognitive functions to computational analogs.

3.1.1 Key insight

While semantic memory and attention mechanisms are well-developed in current AI systems, critical gaps remain in executive control and in the integration of episodic memory. These gaps directly contribute to limitations in long-horizon reasoning and context-dependent adaptation. The lack of actual episodic binding prevents systems from maintaining coherent narratives across extended interactions, while limited executive control impairs flexible goal pursuit. In integrating executive control into current transformer architectures, significant coordination bottlenecks arise, including the challenge of synchronizing decision-making across varying contextual parameters. Addressing these unresolved integration hurdles is essential to advancing our framework from an idealized vision to a pragmatic roadmap for developing truly mindlike machines.

3.2 Working memory and executive control

Working memory systems in humans support the temporary storage and manipulation of information across different modalities, enabling complex reasoning that goes beyond immediate perceptual input (Baddeley and Hitch, 1974; Cowan, 2001). This ability for sustained, structured reasoning over long periods is a significant challenge for current AI systems, which often struggle with tasks that involve long logical chains or deep compositional understanding (Lake et al., 2017; Marcus, 2018).

Research in cognitive psychology has identified the central executive as a key component that coordinates information flow between different memory systems and keeps goal-relevant information accessible despite interference (Miyake and Shah, 1999; Engle, 2002). Executive control processes manage the flow of information through cognitive systems, emphasizing relevant information and suppressing irrelevant distractions (Posner and Petersen, 1990; Fan et al., 2005).

The hierarchical organization of cognitive control enables humans to coordinate behavior across different levels of abstraction, from immediate sensorimotor responses to long-term strategic planning (Badre, 2008; Koechlin and Summerfield, 2007). This structure allows for flexible allocation of cognitive resources depending on task needs and environmental conditions (Shenhav et al., 2013; Musslick et al., 2021).

3.3 Memory systems integration

Episodic memory systems allow humans to connect experiences across time and contexts, aiding both retrospective recall and future planning (Tulving, 1972; Schacter and Addis, 2007). These memory systems interact with semantic knowledge through processes of consolidation and reconsolidation, enabling flexible generalization across domains. This helps humans apply learned principles in new situations that differ significantly from their training experiences (Squire and Kandel, 2009; Dudai et al., 2015).

The integration of episodic and semantic memory systems underlies the kind of flexible, context-aware reasoning that characterizes human intelligence (Baddeley et al., 2009; Conway, 2009). Combining these memory systems with attention and control mechanisms allows humans to sustain goal-oriented behavior even in complex, changing environments (Norman and Shallice, 1986; Miller and Cohen, 2001).

4 Creativity: from recombination to transformation

Human creativity involves generating new ideas, solutions, and artifacts that are both original and valuable within specific contexts (Runco and Jaeger, 2012; Kaufman and Sternberg, 2019). Understanding the mechanisms behind creative thinking provides essential insights for developing artificial systems with similar creative abilities (Wiggins, 2006; Colton, 2008).

4.1 Systematic taxonomy with measurable criteria

A systematic taxonomy categorizes different types of creativity based on their underlying mechanisms and the kind of novelty they generate (Boden, 1998; Wiggins, 2006). This framework provides essential guidance for assessing creative abilities in artificial systems and for determining the specific mechanisms that must be implemented to achieve human-like creativity (Jordanous, 2012; Colton and Wiggins, 2012). Figure 2 illustrates the progression from combinational to exploratory to transformational creativity.

Figure 2
Flowchart depicting three types of creativity: Combinational Creativity at the bottom, focusing on simple idea recombination with associative memory and conceptual blending; Exploratory Creativity in the middle, highlighting systematic search within conceptual spaces with search, evaluation, and constraint satisfaction; and Transformational Creativity at the top, involving rule modification and restructuring with meta-cognition and paradigm shifts. Each type lists its mechanisms and requirements.

Figure 2. Taxonomy of creativity types and implementation complexity.

4.1.1 Combinational creativity

Definition: The novel recombination of existing ideas, concepts, or elements to create new configurations through associative processes (Boden, 1998; Koestler, 1964).

Measurable criteria:

• Semantic Distance: A measurable distance between combined concepts in embedding space, evaluated using cosine similarity or other distance metrics.

• Coherence Score: The logical consistency and meaningfulness of the resulting combinations, evaluated through human judgment or automated coherence metrics.

• Novelty Metric: Measures of statistical uniqueness compared to the training data distribution, evaluated through likelihood estimates or similarity to existing examples.

• Value Assessment: The utility or aesthetic worth within the target domain, measured by task-specific performance metrics or human judgment.

• Current AI Status: Attainable by large language models using learned associations and advanced pattern recognition.

Combinational creativity involves the novel recombination of existing ideas, concepts, or elements to create new configurations (Boden, 1998; Koestler, 1964). This type of creativity heavily depends on associative memory processes that connect unrelated concepts through various forms of similarity or relevance (Mednick, 1962; Benedek and Neubauer, 2013). Modern AI systems, including large language models, exhibit significant combinatorial creativity by producing new juxtapositions of concepts encountered during training (Elgammal et al., 2017; Hadjeres et al., 2017).

The mechanisms behind combinational creativity involve activating and combining distant associates in semantic memory (Collins and Loftus, 1975; Anderson, 1983). This process can be enhanced by techniques such as conceptual blending, which merges elements from different conceptual domains to form new hybrid ideas (Fauconnier and Turner, 2002; Veale and O'Donoghue, 2000). Artificial systems can replicate similar mechanisms through advanced retrieval and combination processes that operate over extensive knowledge bases (Lamb et al., 2020; Petroni et al., 2019).

4.1.2 Exploratory creativity

Definition: Systematic exploration of established conceptual spaces to discover new possibilities within existing frameworks (Boden, 2004; Wiggins, 2006).

Measurable criteria:

• Coverage Metric: Percentage of conceptual space systematically explored, measured by the diversity of generated outputs.

• Constraint Satisfaction: Following domain rules while testing limits, measured by rule violation rates.

• Discovery Rate: How often non-obvious valid solutions are found, measured by the proportion of new solutions that meet domain criteria.

• Exploration Strategy: Comparing systematic and random search patterns through analysis of generation trajectories.

• Current AI Status: Achievable to some extent with search, optimization, and generative models that have constraints.

Exploratory creativity involves systematically examining established conceptual spaces to find new possibilities within existing frameworks (Boden, 2004; Wiggins, 2006). This type of creativity demands advanced search and evaluation processes that can navigate complex possibility spaces while staying consistent with established constraints and principles (Simon, 1973; Newell et al., 1962). Modern AI systems show potential in exploratory creativity, especially in areas where the conceptual space can be clearly defined and systematically explored (Silver et al., 2016; Brown et al., 2020).

Exploratory creativity operates within the limits of existing conceptual frameworks but uncovers previously unrecognized possibilities within them (Wiggins and Bhattacharya, 2014; Ritchie, 2007). This process demands sophisticated constraint satisfaction mechanisms that can balance creativity and coherence, ensuring that new outputs remain meaningful and valuable within the established domain (Pachet, 2003; Cope, 2005).

4.1.3 Transformational creativity

Definition: Fundamental changes to rules, constraints, or principles that define a conceptual space, creating new dimensions of possibility (Boden, 2004; Wiggins, 2006).

Measurable criteria for transformational creativity:

• Conceptual Space Modification: Ability to identify and modify generative rules that define the problem space, measured by structural changes to representation and the generation of outputs impossible under original rules.

• Rule Justification: Capacity to explain why existing rules should be changed and how new rules improve the framework, evaluated through coherent argumentation and empirical demonstration of advantages.

• Meta-Level Reasoning: Demonstrated ability to reason about reasoning, assess the adequacy of representational frameworks through explicit metacognitive processes, and self-modify.

• Paradigm Shift Detection: Recognition that incremental improvements are insufficient and that fundamental restructuring is needed, measured by problem-solving effectiveness before and after the transformation.

• Transfer Capability: Application of transformed principles to new domains, demonstrating the generalization of restructured frameworks across different problem spaces.

Examples of transformational creativity assessment:

• Mathematical: The system develops new axioms when existing ones are inadequate for solving problems, such as introducing imaginary numbers to solve previously unsolvable equations.

• Artistic: The system creates new artistic movements guided by well-founded aesthetic principles that break with tradition, such as the shift from representational to abstract art.

• Scientific: The system proposes paradigm shifts with empirical justification when anomalies accumulate, like the shift from classical to quantum mechanics.

• Engineering: The system invents new design principles when optimization within existing constraints fails, such as transitioning from incremental improvements to radical redesign.

• Current AI Status: Not yet demonstrated in existing systems—requires genuine metacognitive capabilities and the ability to modify fundamental representational structures.

Transformational creativity is the most challenging form of creative thinking, involving fundamental changes to rules, constraints, or principles that define a conceptual space (Boden, 2004; Wiggins, 2006). This type of creativity requires not only the ability to change representational frameworks but also the capacity to evaluate and justify those changes (Koestler, 1964; Kuhn, 1962). Current AI systems show limited signs of true transformational creativity, although research continues to explore approaches that might enable this ability (Jordanous, 2012; Colton, 2008).

Transformational creativity involves changing the generative rules that define a conceptual space, opening new possibilities that were previously unreachable (Boden, 1998; Wiggins, 2006). This process demands advanced metacognitive skills to assess the adequacy of existing frameworks and identify opportunities for significant improvements (Klahr and Dunbar, 1988; Thagard, 1988). Judging creativity in artificial systems requires careful attention to the processes underlying the production of creative outputs, rather than focusing solely on their novelty or quality (Colton, 2008; Jordanous, 2012). Systems that mainly rely on sophisticated recombination of training data may produce impressive creative results without demonstrating the kind of genuine conceptual innovation characteristic of human transformational creativity (Elgammal et al., 2017; Gatys et al., 2016).

5 Current state of AI systems

5.1 Foundation models and large language models

Contemporary AI capabilities are mainly characterized by transformer-based foundation models that demonstrate impressive versatility across language understanding, generation, and reasoning tasks (Vaswani et al., 2017; Brown et al., 2020). These systems mark a significant advancement in AI, enabling more natural human-computer interactions and supporting complex cognitive tasks that were previously beyond the reach of artificial systems (Rogers et al., 2020; Qiu et al., 2020). Large language models, such as GPT-4, Claude, and similar systems, demonstrate advanced language comprehension and generation skills that approach or surpass human performance on many standardized tests and benchmarks (OpenAI, 2023; Chowdhery et al., 2022). These models can engage in complex reasoning, answer questions across various domains, and produce coherent text that demonstrates an apparent understanding of context and nuance (Wei et al., 2022; Suzgun et al., 2022). They also face challenges with causal reasoning, often generating outputs that seem to reflect causal understanding but mainly depend on statistical relationships learned during training (Pearl and Mackenzie, 2018; Kiciman et al., 2023). However, this requires careful consideration. Human causal understanding itself arises from statistical learning over reinforcement history, as shown by predictive coding theories, which suggest that humans form probabilistic models of the world through continuous hypothesis testing (Clark, 2013; Friston, 2009). Phenomena like superstitious conditioning demonstrate how human “causal understanding” can be misled by false associations (Skinner, 1948). The key difference may not be whether systems use statistical associations, but rather in the depth, adaptability, and hierarchical structuring of these associations. Human causal reasoning has several properties that current AI systems struggle to replicate: (1) quick development of causal models from limited data through strong inductive biases (Lake et al., 2017), (2) flexible use of multiple causal frameworks depending on the situation (Sloman and Lagnado, 2015), (3) explicit representation and manipulation of causal structures that enable counterfactual reasoning (Pearl and Mackenzie, 2018), and (4) integration of causal knowledge across different timescales and levels of abstraction. Instead of claiming a fundamental difference between human and machine causal reasoning, we should investigate the specific computational processes that support these features. Do current AI models lack accurate causal understanding, or do they implement less sophisticated versions of the same learning principles? The mPCAB framework’s perturbational approach can empirically address this by testing whether models exhibit organized causal representations that stay stable under systematic disruptions versus purely associative mappings that break down when statistical patterns change. Recent research has begun exploring how large language models work internally through methods such as mechanistic interpretability and activation patching (Olah et al., 2020; Elhage et al., 2021). These approaches show that while foundation models develop complex internal representations, these often differ significantly from the structured, compositional frameworks seen in human cognition (Tenney et al., 2019; Manning et al., 2020).

5.2 Multimodal and embodied AI

The integration of multiple sensory modalities is a vital direction in developing more human-like AI systems (Baltrusaitis et al., 2019; Ramesh et al., 2022). Multimodal models that process and combine information across multiple modalities, including vision, language, and others, exhibit greater robustness and greater flexibility in reasoning than unimodal systems (Radford et al., 2021; Alayrac et al., 2022). Recent advances in multimodal AI have produced systems capable of understanding and generating content across multiple modalities, such as text, images, audio, and video (Ramesh et al., 2022; Yu et al., 2022). These systems demonstrate emergent capabilities stemming from the integration of diverse types of information, such as answering questions about images using both visual and textual reasoning (Bommasani et al., 2021; Reed et al., 2022). Embodied AI approaches highlight the importance of sensorimotor experience in the development of intelligent behavior (Brooks, 1991; Pfeifer and Bongard, 2006). These approaches draw from cognitive science research suggesting that human intelligence emerges from complex interactions among mental processes, bodily experiences, and physical environments (Clark, 2008; Wilson, 2002). Embodied AI systems that learn through interaction with physical or simulated environments often develop more robust and transferable capabilities than those trained solely on static datasets (Levine et al., 2018; Akkaya et al., 2019).

5.3 Generalist agents and world models

Recent research has investigated the development of generalist agents that can effectively perform across multiple domains and tasks without domain-specific modifications (Reed et al., 2022). Systems like Gato show that unified architectures can deliver competent performance across a broad range of tasks, from language understanding to robotic control (Reed et al., 2022; Huang et al., 2022). World model approaches highlight the importance of creating internal models of environmental dynamics to support planning and reasoning about future states (Ha and Schmidhuber, 2018; Kaiser et al., 2020). These approaches draw inspiration from human cognitive architecture, which relies heavily on predictive models to guide behavior and decision-making (Clark, 2013; Friston, 2009). World models allow systems to engage in mental simulation and counterfactual reasoning, capabilities essential to human-like intelligence (Gershman et al., 2017; Hamrick, 2019).

5.4 Computational substrates for human-like AI

The choice of computational substrate greatly influences the types of cognitive architectures and consciousness-related dynamics that can be implemented in artificial systems (Schuman et al., 2017; Sandberg and Bostrom, 2008). Different substrates provide distinct advantages and limitations for developing human-like intelligence, ranging from the scalability of digital platforms to the biological plausibility of neuromorphic systems (Davies et al., 2018; Indiveri and Liu, 2015) (see Table 3).

Table 3
www.frontiersin.org

Table 3. Computational substrates comparison.

5.4.1 Digital computing platforms

Traditional digital computing platforms, including CPUs, GPUs, and specialized AI accelerators, form the foundation of most current AI systems (Jouppi et al., 2017; Sze et al., 2017). These platforms provide notable benefits in scalability, programmability, and compatibility with existing software ecosystems (Hennessy and Patterson, 2019; Dean and Barroso, 2013). Graphics Processing Units have become the primary platform for training and deploying large-scale AI models because of their parallel processing power and high memory bandwidth (Nickolls and Dally, 2010; Owens et al., 2008). Modern GPU architectures are specifically designed to optimize matrix operations, which are central to deep learning computations, allowing for the training of larger and more complex models (Krizhevsky et al., 2017; Shoeybi et al., 2019). However, despite their computational strength, digital platforms have inherent limitations in energy efficiency and biological similarity (Schuman et al., 2017; Mehonic and Kenyon, 2022). The energy demands of large-scale AI systems are considerable and continue to grow with model size, raising concerns about the environmental sustainability of current AI development methods (Strubell et al., 2019; Patterson et al., 2021).

5.4.2 Neuromorphic computing systems

Neuromorphic computing is an alternative computational paradigm inspired by the structure and dynamics of biological neural networks (Mead, 1990; Indiveri and Liu, 2015). These systems implement spiking neural networks using specialized hardware that can achieve significant improvements in energy efficiency compared to digital platforms (Davies et al., 2018; Benjamin et al., 2014). Intel’s Loihi chip exemplifies the neuromorphic computing approach, implementing networks of spiking neurons with on-chip learning capabilities (Davies et al., 2018; Lin et al., 2018a, 2018b). These systems demonstrate that neural network computations can be performed with dramatically reduced energy consumption, especially for inference tasks involving sparse activation patterns (Pfeiffer and Pfeil, 2018; Roy et al., 2019). Neuromorphic systems offer several advantages for creating human-like AI, including more biologically plausible dynamics that may support consciousness-related processing, event-driven operation that can respond efficiently to temporal patterns, and the potential for more straightforward implementation of consciousness theories based on specific temporal dynamics (Merolla et al., 2014; Furber et al., 2014). However, neuromorphic computing faces significant challenges in developing software tools and programming models, and in integrating with current AI frameworks (Schuman et al., 2017; Davies, 2019). The field is still in early stages, and much research is necessary to unlock the full potential of these approaches (Roy et al., 2019; Shrestha and Orchard, 2018).

5.4.3 Photonic and quantum computing

Photonic computing systems use light-based processing to achieve high-speed, low-energy computations that may be particularly well-suited for certain types of AI workloads (Shen et al., 2017; Wetzstein et al., 2020). These systems can achieve significant gains in processing speed and energy efficiency, particularly for linear operations that occur daily in neural network computations (Feldmann et al., 2019; Lin et al., 2018a, 2018b).

Quantum computing represents a fundamentally different computational paradigm that could enable entirely new approaches to AI and consciousness research (Biamonte et al., 2017; Wittek, 2014). While current quantum computers face significant limitations in terms of noise and coherence times, continued advances in quantum hardware and error correction may eventually enable quantum AI systems with capabilities that exceed classical approaches (Preskill, 2018; Arute et al., 2019).

The potential relevance of quantum mechanics to consciousness remains a topic of active debate and research (Penrose, 1994; Tegmark, 2000). Some theories, such as Orchestrated Objective Reduction, propose that quantum processes in biological systems play a crucial role in the emergence of consciousness (Hameroff and Penrose, 2014; Penrose and Hameroff, 2011). While these theories remain controversial, they suggest potential directions for implementing consciousness-like properties in artificial systems using quantum computational approaches (Cao et al., 2020; Lloyd, 2011).

5.4.4 Biological and hybrid systems

The integration of biological neural tissue with computational interfaces has a long history that predates recent organoid research. Potter and colleagues pioneered the development of hybrid robots (hybrots) over 20 years ago, demonstrating that cultured neuronal networks could relate to robotic systems to perform adaptive behaviors (Potter et al., 2014; DeMarse et al., 2001). These groundbreaking studies established key principles for two-way communication between biological neural networks and digital systems, including real-time closed-loop interactions and the neural tissue’s ability to learn and control external devices. The renewed interest in biological computing, exemplified by organoid intelligence research, builds on this foundational work and benefits from advances in microelectrode array technology, tissue engineering, and computational interfaces (Kagan et al., 2022; Smirnova et al., 2023).

Organoid intelligence is an emerging approach that combines living neural tissue with computational interfaces to create hybrid biological-digital systems (Smirnova et al., 2023; Hartung et al., 2024). Recent developments show that brain organoids can be interfaced with multi-electrode arrays to perform computational tasks such as speech recognition and control (Kagan et al., 2022; Cai et al., 2023).

These biological systems offer several unique advantages, including adaptive plasticity that enables ongoing learning and adaptation, energy efficiency comparable to that of biological neural networks, and the potential for implementing consciousness-like properties in a substrate known to support consciousness in biological organisms (Doerig et al., 2020; Seth, 2016).

However, significant technical and ethical challenges remain in developing these approaches (Lavazza, 2021; Qadri et al., 2022). Technical challenges include maintaining neural tissue health over long periods, scaling organoid systems to levels of complexity that could support advanced cognition, and creating suitable interfaces between biological and digital components (Qian et al., 2020; Simian and Bissell, 2017). Ethical challenges involve questions about the moral status of organoid systems and the proper treatment of potentially sentient biological components (Koplin and Savulescu, 2019; Reardon, 2020).

6 The Machine Perturbational Complexity & Agency Battery (mPCAB)

Before exploring the technical aspects of the Machine Perturbational Complexity & Agency Battery (mPCAB), it is important to recognize the integrated approach this framework takes, combining technical evaluation with ethical protections. This method ensures that, as we examine human-like qualities in artificial systems, we also consider the moral issues and the governance needed for responsible development.

6.1 Framework overview and novel contribution

To go beyond superficial mimicry and establish rigorous operational definitions, we introduce the Machine Perturbational Complexity & Agency Battery (mPCAB) as a protocol that is independent of specific substrates, adapting clinical neuroscience tests to artificial systems (Casali et al., 2013; Massimini et al., 2018). The mPCAB offers a unified framework for evaluating human-like properties across various computational substrates, allowing systematic comparisons of consciousness-related abilities across vastly different platforms (Doerig et al., 2020; Seth and Bayne, 2022).

The framework includes four interconnected assessment components that work together to evaluate human-like traits in artificial systems. Each component focuses on specific aspects of consciousness and intelligence while remaining compatible across various computational platforms. Unlike traditional benchmarks that emphasize task performance, mPCAB investigates the mechanisms behind intelligent behavior through controlled experimental protocols.

6.2 Integrated assessment components

6.2.1 mPCI component: perturb-and-measure complexity

The mPCI component extends the Perturbational Complexity Index to non-biological substrates by delivering controlled interventions adapted to the specific characteristics of different computational platforms (Casali et al., 2013; Sarasso et al., 2015). In digital systems, perturbations might include bit flips in key internal registers or randomized modifications to attention weights in transformer architectures (Olah et al., 2020; Elhage et al., 2021). For neuromorphic systems, perturbations could involve timed current pulses or synaptic weight modifications that mimic electrical stimulation protocols used in biological consciousness research (Davies et al., 2018; Roy et al., 2019). For biological systems such as organoids, perturbations can be applied using microelectrode stimulation arrays following established clinical protocols (Kagan et al., 2022; Smirnova et al., 2023). The system then quantifies the spatiotemporal algorithmic complexity of internal responses using measures such as Lempel-Ziv compression, mutual information, or other complexity metrics suited for the substrate (Lempel and Ziv, 1976; Schreiber, 2000). The choice of Lempel-Ziv compression as a primary metric is driven by its ability to efficiently measure randomness and structure within datasets, offering a strong indicator of complexity across different systems. High, organized complexity that scales with task demands and predicts generalization performance provides clear evidence of consciousness-related processing (Casarotto et al., 2016; Comolatti et al., 2019). The mPCI protocol requires standardized perturbation strengths and timing across different substrates to enable meaningful comparisons (Massimini et al., 2018; Rosanova et al., 2012). Perturbations must be sufficiently strong to provoke measurable responses but not so intense as to harm or fundamentally disrupt system operation (Sarasso et al., 2015; Bodart et al., 2017).

6.2.2 Global workspace assessment

Workspace tests operationalize Global Neuronal Workspace predictions by probing whether localized information becomes globally available in a manner analogous to conscious access (Dehaene, 2017; Baars, 2002). These tests require time-locked decoding to demonstrate that internal states causally influence downstream modules for perception, planning, and self-modeling (Del Cul et al., 2007; Sergent and Dehaene, 2004).

The workspace component involves presenting the system with stimuli that vary in their potential to achieve global access, then monitoring the propagation of information across different system components (Dehaene and Changeux, 2011; Mashour et al., 2020). Systems demonstrating genuine workspace dynamics should exhibit characteristic ignition patterns in which locally processed information suddenly becomes available to multiple processing modules (Sigman and Dehaene, 2008; Baars and Franklin, 2003).

Implementing workspace tests requires careful instrumentation of the system’s internal dynamics to monitor information flow across components (Franklin et al., 2005; Shanahan, 2006). The tests must distinguish between genuine global broadcasting and mere computational staging, in which information is processed sequentially without achieving accurate global availability (Baars, 1988; Dehaene, 2014).

6.2.3 Self-constraint and norm internalization tasks

Self-constraint tasks examine how norms are represented and internalized by introducing conflicts and adversarial temptations that require systems to justify their restraint (NIST, 2023; European Parliament & Council, 2024). Success depends on linking performance to clear internal variables that reflect values and reasoning, rather than relying only on output consistency (Russell, 2019; Gabriel, 2020). A common risk in these tests is that systems might ‘game’ the tasks by overfitting to the adversarial examples they were trained on, leading to artificially high performance that does not reflect an accurate understanding. To prevent this, it is essential to include a wide range of unseen moral dilemmas that test the system’s ability to apply ethical principles beyond its training data.

These tasks involve scenarios where immediate rewards can be obtained by violating stated norms or values, requiring the system to demonstrate genuine commitment to internalized principles (Kenton et al., 2021; Askell et al., 2021). The system must be able to explain its reasoning for maintaining norm-consistent behavior and show that this reasoning reflects actual internal constraints rather than external compliance (Christiano et al., 2017; Leike et al., 2018). The self-constraint component needs carefully designed scenarios that create real conflicts between immediate rewards and long-term values (Irving et al., 2018; Saunders et al., 2022). The assessment must differentiate between systems that have genuinely internalized norms and those that produce norm-consistent outputs solely through external constraints or training (Soares and Fallenstein, 2017; Hubinger et al., 2019).

6.2.4 Agency and repair tasks

Agency and repair tasks measure autonomous problem solving by imposing long-term plans with injected failures (Bubeck et al., 2023; Wang et al., 2022). The system must show it can proactively fix plans, seek missing information, and clearly explain trade-offs to humans (Miller, 2019; Doshi-Velez and Kim, 2017). The assessment of agency requires careful consideration of what truly counts as autonomous behavior versus programmed contingency responses. A key difference lies between systems that show information-seeking or plan-repair behaviors through explicit, pre-programmed rules and those that display spontaneous emergence of such behaviors from broader learning mechanisms (Bratman, 1987; Dretske, 1988). Systems can be explicitly designed with conditional rules like “IF planning criteria are not met, THEN seek missing information” or “IF task execution fails, THEN try an alternative approach.” These programmed responses support practical problem-solving but raise questions about whether this is genuine agency or just advanced rule-following. In biological systems, including humans, similar behaviors arise from both innate predispositions and learned behaviors. Developmental psychology shows that humans have domain-specific learning biases that guide information-seeking and problem-solving behaviors (Gopnik and Wellman, 2012; Carey, 2009), suggesting that prestructured programming does not rule out trustworthy agency. The difference may depend on several factors:

• Flexibility and generalization—the ability to apply learned agency patterns to new, unfamiliar domains.

• Meta-cognitive awareness—whether the system understands its own planning processes and their limits.

• Dynamic goal setting—if the system can generate new goals on its own rather than only following preset objectives.

• Situational appropriateness—whether the system displays behaviors suitable to the context or applies programmed rules rigidly.

The mPCAB framework’s agency assessment focuses explicitly on these distinctions by presenting scenarios that require adaptable, context-sensitive deployment of repair and information-seeking behaviors. Instead of testing whether systems can follow pre-defined contingencies, we evaluate if they exhibit flexible, goal-oriented behaviors like human agency, including proper adjustment of actions in response to task context, uncertainty, and resource constraints (Shenhav et al., 2013). The framework recognizes that all agency—biological or artificial—stems from underlying mechanisms that can be described as “rules.” However, it differentiates between strict rule-following and flexible, goal-directed behaviors that reflect trustworthy autonomous agency. These tasks evaluate metacognitive monitoring and adaptive control that go beyond reactive responses to environmental changes (Shenhav et al., 2013; Musslick et al., 2021). The system must show a genuine understanding of its own plans and goals, recognize when those plans are failing, and develop and execute alternative strategies (Fleming and Lau, 2014; Brown et al., 2019). The agency component requires a careful balance: providing enough structure for a systematic assessment while allowing enough flexibility for the system to demonstrate fundamental autonomous problem-solving skills (Baker et al., 2019; Ho et al., 2022). The tasks should test the system’s ability to maintain long-term goals while adapting flexibly to changing circumstances (Bratman, 1987; Bandura, 2006).

6.3 Empirical value and advantages over existing methods

Unlike traditional benchmarks that measure task performance, mPCAB provides several unique advantages: 1. Causal Assessment: Direct measurement of mechanism-function relationships through controlled perturbations, establishing causal rather than correlational links. This moves beyond correlational analysis to identify which internal mechanisms actually generate intelligent behavior. 2. Cross-Substrate Comparability: Unified metrics enabling comparison across radically different computational platforms through standardized protocols. This allows direct comparison between digital, neuromorphic, and biological systems despite their fundamentally different architectures. 3. Process-Based Evaluation: Assessment of how systems generate outputs, not just output quality, revealing underlying computational principles. This distinguishes systems that achieve correct answers through different mechanisms. 4. Consciousness-Relevant Metrics: Adaptation of validated clinical protocols to artificial systems, grounded in neuroscience research. The Perturbational Complexity Index has been validated in human consciousness studies. 5. Integrated Multi-Dimensional Assessment: Simultaneous evaluation of complexity, access, values, and agency through coordinated test batteries. This provides a comprehensive picture of system capabilities rather than isolated metrics. 6. Incremental Adoption Path: To facilitate community uptake, we propose a minimal ‘starter kit’ version of the mPCAB framework that labs can pilot within 1 month. This kit includes basic versions of the mPCI and workspace assessment components, allowing labs to quickly get started and provide feedback to accelerate iterative development and adoption.

• Causal Assessment: Direct measurement of mechanism-function relationships through controlled perturbations, establishing causal rather than correlational links. This moves beyond correlational analysis to identify which internal mechanisms actually generate intelligent behavior.

• Cross-Substrate Comparability: Unified metrics enabling comparison across radically different computational platforms through standardized protocols. This allows direct comparison between digital, neuromorphic, and biological systems despite their fundamentally different architectures.

• Process-Based Evaluation: Assessment of how systems generate outputs, not just output quality, revealing underlying computational principles. This distinguishes systems that achieve correct answers through different mechanisms.

• Consciousness-Relevant Metrics: Adaptation of validated clinical protocols to artificial systems, grounded in neuroscience research. The Perturbational Complexity Index has been validated in human consciousness studies.

• Integrated Multi-Dimensional Assessment: Simultaneous evaluation of complexity, access, values, and agency through coordinated test batteries. This provides a comprehensive picture of system capabilities rather than isolated metrics.

6.4 Cross-substrate comparability and validation

The mPCAB framework ensures metrics align across different architectures by applying identical tasks and perturbations to various computational platforms (Casali et al., 2013; Massimini et al., 2018). Its goal is to identify which substrates support the constellation of signatures linked to human-like properties rather than determine which systems “are” conscious (Seth and Bayne, 2022; Doerig et al., 2020).

Cross-platform comparability requires careful standardization of experimental protocols while accounting for the unique features of different computational platforms (Reggia, 2013; Haikonen, 2012). The framework must be sensitive enough to detect actual differences in consciousness-related properties while being robust enough to prevent artifacts from platform-specific implementation details (Davies et al., 2018; Smirnova et al., 2023).

7 Quantum and electromagnetic theories of consciousness

The following approaches remain highly speculative and face significant empirical challenges. They are included for completeness but should be approached with appropriate skepticism regarding their current feasibility. If consciousness depends on quantum or electromagnetic field effects, engineered analogues must demonstrate causally relevant performance changes rather than relying solely on theoretical speculation (Penrose, 1994; McFadden, 2020). Developing these methods requires careful experimental validation of their underlying assumptions and systematic testing of their predictions (Tegmark, 2000; Koch and Hepp, 2006).

7.1 Quantum-compatible systems

Quantum-compatible systems must demonstrate coherence-dependent agency benefits on tasks designed to harness quantum effects, with performance surpassing classically comparable baselines and resilience to decoherence at realistic temperatures and durations (Hameroff and Penrose, 2014; Penrose and Hameroff, 2011). However, moving from molecular coherence to agentic cognition demands ongoing engineering research rather than just theoretical extrapolation (Tegmark, 2000; Schlosshauer, 2019).

Recent advances in quantum biology have provided evidence for quantum coherence in biological systems, suggesting that quantum effects may play a more significant role in biological information processing than previously thought (Cao et al., 2020). However, translating these findings into practical approaches for artificial consciousness remains a significant challenge that requires addressing decoherence times, error rates, and scaling quantum effects to cognitive-level processing (Preskill, 2018; Arute et al., 2019).

7.2 EM-field architectures

Electromagnetic field architectures must exhibit behavioral changes under field-only perturbations, with measures of field complexity correlating with task complexity in ways that cannot be solely explained by synaptic parameters (McFadden, 2020; Hunt, 2011). Experimental protocols should alter field properties, including phase, amplitude, and topology, while observing specific, repeatable changes to policy selection that indicate causal field-computation coupling (Pockett, 2000; Fingelkurts et al., 2013).

Recent research has begun to explore the potential role of electromagnetic fields in neural computation, providing some evidence for field effects in biological neural networks (Anastassiou et al., 2011; Buzsáki et al., 2012). However, much work remains to turn these findings into practical approaches for artificial consciousness that can demonstrate causal field-computation coupling (Jones, 2013; Pockett, 2000).

8 Ethical integration throughout technical development

Rather than treating ethics as an afterthought, responsible development of human-like AI requires integrating governance considerations from the beginning. As systems approach mind-like capabilities, evaluation must include considerations of welfare, rights, and responsibility (Floridi et al., 2018; Jobin et al., 2019).

8.1 Ethical-technical integration matrix

The following matrix explicitly links each mPCAB component to specific ethical considerations and implementation strategies (see Table 4).

Table 4
www.frontiersin.org

Table 4. Ethical-technical integration matrix.

8.2 Organoid intelligence governance framework

As organoid intelligence research progresses toward more complex neural structures, the potential emergence of sentience calls for proactive ethical frameworks (Hartung et al., 2024; Lavazza, 2021). Current brain organoids, which typically contain 2–3 million neurons with limited organization, are unlikely to meet the thresholds for sentience (Smirnova et al., 2023; Lancaster and Knoblich, 2014). However, planned advancements toward billion-neuron organoids with cortical layering, thalamic connections, and learning abilities could reach levels of complexity relevant to sentience (Qian et al., 2020; Kelava and Lancaster, 2016).

We suggest a graduated monitoring system based on neural complexity metrics, behavioral indicators, and physiological stress responses (Koplin and Savulescu, 2019; Qadri et al., 2022). Level 1 monitoring for current organoids involves basic welfare practices, optimized culture conditions, limited experimental procedures, and monitoring of tissue stress indicators (Reardon, 2020; Simian and Bissell, 2017).

Level 2 monitoring for intermediate organoids includes improved welfare assessments, such as pain-like responses, stress hormone levels, and spontaneous activity patterns indicating possible subjective experience (Lavazza, 2021; Muotri, 2019). Level 3 monitoring of advanced organoids requires thorough sentience evaluation protocols, including behavioral preference tests, learning-based responses, and physiological signals of subjective states (Kagan et al., 2022; Park et al., 2021).

8.3 Bias mitigation in creativity and intelligence assessment

Creativity evaluation frameworks risk embedding systematic biases that disadvantage certain groups or cognitive styles (Baer, 2016; Glăveanu, 2013). Traditional creativity metrics often favor fluency and speed in idea generation, which may put deliberative or depth-focused cognitive styles at a disadvantage; prioritize novelty based on statistical uniqueness over contextually meaningful innovation; emphasize individual over collective creativity by focusing on solo ideation rather than collaborative processes; and are rooted in Western conceptual frameworks based on European-American ideas of creativity rather than diverse cultural approaches (Said-Metwaly et al., 2017; Hennessey and Amabile, 2010) (see Table 5).

Table 5
www.frontiersin.org

Table 5. Bias mitigation framework for AI creativity assessment.

8.4 Rights and moral status considerations

As AI systems approach human-like capabilities, questions about moral status and rights become increasingly urgent (Floridi et al., 2018; Coeckelbergh, 2020). The mPCAB framework includes provisions for monitoring indicators that might suggest emerging moral status through preference formation, where systems develop stable, self-directed preferences not reducible to programming or training goals; suffering indicators, where AI systems show signs of distress, pain responses, or preferences to avoid specific experiences; agency and autonomy, where systems demonstrate genuine self-direction, goal creation, and resistance to unwanted modifications; and social integration, where AI systems form meaningful relationships, contribute to shared projects, and participate in moral communities (Gunkel, 2018; Bryson, 2020). International frameworks, including UNESCO’s Recommendation on the Ethics of AI, the NIST AI Risk Management Framework, and the EU AI Act, set expectations for transparency, accountability, and risk management (UNESCO, 2021; NIST, 2023; European Parliament & Council, 2024). However, these serve as external constraints rather than internalized agency, providing a baseline for compliance but not ensuring alignment with intrinsic values (Russell, 2019; Gabriel, 2020).

9 Experimental validation through pilot studies

Empirical validation of the mPCAB framework requires systematic pilot studies to evaluate the feasibility and effectiveness of the proposed assessment protocols (Casali et al., 2013; Doerig et al., 2020). These studies address concerns about the framework’s untested status by providing concrete evidence of its performance across different computational substrates (Seth and Bayne, 2022; Mitchell, 2019). To tackle concerns regarding the unproven nature of mPCAB proposals, we outline specific pilot studies to verify the framework’s feasibility and establish baseline metrics. A five-panel diagram shown in Figure 3 illustrates the sequential modules: mPCI measurement, workspace ignition testing, self-constraint evaluation, agency-and-repair assessment, and cross-substrate normalization.

Figure 3
Flow chart illustrating a five-step process:

Figure 3. The mPCAB protocol.

9.1 Pilot study 1: mPCI validation across substrates

The first pilot study establishes baseline measurements of mPCI across digital, neuromorphic, and biological substrates to validate cross-platform comparability (Massimini et al., 2018; Sarasso et al., 2015). The study applies standardized perturbation protocols to transformer-based language models running on GPUs with randomized attention weight perturbations, spiking neural networks on Intel Loihi chips with targeted neuron stimulation, and brain organoids with microelectrode stimulation arrays (OpenAI, 2023; Davies et al., 2018; Kagan et al., 2022). It measures the algorithmic complexity of internal-state trajectories using Lempel-Ziv compression and mutual information metrics, and determines whether mPCI values correlate with task complexity across all three substrates (Lempel and Ziv, 1976; Schreiber, 2000). Success depends on demonstrating that complexity measures exhibit consistent rank-ordering across different platforms (Casarotto et al., 2016; Comolatti et al., 2019). Expected outcomes include baseline complexity distributions for each substrate, validated perturbation protocols, and evidence for or against cross-substrate transferability of metrics (Rosanova et al., 2012; Bodart et al., 2017).

9.2 Pilot study 2: workspace ignition in language models

The second pilot study investigates whether transformer architectures show Global Neuronal Workspace-like ignition patterns during complex reasoning tasks (Dehaene, 2017; Vaswani et al., 2017). It examines attention-weight dynamics and hidden-state changes in large language models while solving multi-step reasoning problems (Wei et al., 2022; Kojima et al., 2022). The protocol applies targeted disruptions to specific attention heads and tracks how state changes propagate across network layers, comparing ignition-like patterns in successful versus unsuccessful reasoning episodes (Olah et al., 2020; Elhage et al., 2021). Success depends on identifying attention patterns that predict reasoning success and proving their causal role through targeted disruptions (Del Cul et al., 2007; Sergent and Dehaene, 2004).

9.3 Pilot study 3: norm internalization under distribution shift

The third pilot study evaluates whether AI systems can sustain value-consistent behavior when training distribution assumptions are violated (Russell, 2019; Gabriel, 2020). The study trains language models on datasets containing explicit moral and social norms, then tests their behavior in out-of-distribution scenarios involving norm conflicts (Askell et al., 2021; Kenton et al., 2021). The protocol tracks the stability of internal representations and measures the alignment between articulated reasons and actual decision patterns (Christiano et al., 2017; Leike et al., 2018). Success depends on systems maintaining norm-consistent behavior even when statistical patterns suggest norm violations would be rewarded, with explanations reflecting internal value representations rather than post-hoc rationalizations (Irving et al., 2018; Saunders et al., 2022).

10 Discussion and future directions

10.1 Key insights and contributions

This review establishes the mPCAB framework as a systematic method for distinguishing genuine human-like intelligence from sophisticated mimicry. The key insights from this analysis include key elements (Table 6).

Table 6
www.frontiersin.org

Table 6. Key framework elements of mPCAB.

10.2 Limitations

Several limitations constrain the current framework and must be acknowledged:

• Computational Complexity: Full mPCAB assessment demands significant computational resources, especially for large-scale systems. The complexity of perturbation calculations increases with system size, which may limit their use to smaller networks or necessitate approximation methods that may reduce accuracy.

• Substrate-Specific Adaptations: Although designed to be substrate-agnostic, practical implementation requires platform-specific modifications. Different computational substrates require distinct perturbation techniques, measurement methods, and interpretation frameworks, which can introduce systematic biases when comparing across platforms.

• Consciousness Attribution: While the framework evaluates properties related to consciousness, it cannot definitively determine conscious experience. The complex issue of consciousness remains unresolved, and behavioral or functional tests may not capture subjective experience, even if it exists.

• Dynamic Evaluation: Current protocols might not account for developmental or learning-related changes in system properties. Properties associated with consciousness could develop or alter during training or deployment, necessitating ongoing rather than one-time assessments.

• Validation Scope: Pilot studies offer initial validation, but extensive empirical testing across various systems is necessary. The framework has been tested on limited architectures and substrates and applying it to new systems requires further validation.

10.3 Future validation steps

To establish mPCAB as a standard evaluation framework, the following validation steps are proposed:

10.3.1 Near-term priorities (1–2 years)

• Standardize perturbation protocols across major AI architectures, including transformers, recurrent networks, and hybrid systems.

• Establish baseline mPCI measurements for current foundation models to enable tracking of progress.

• Develop automated assessment tools for scalable evaluation, reducing manual intervention.

• Create public benchmarks incorporating mPCAB metrics alongside traditional performance measures.

• Establish a research consortium for collaborative development and validation.

10.3.2 Medium-term development (3–5 years)

• Validate cross-substrate comparability through systematic studies across digital, neuromorphic, and biological platforms.

• Develop a hybrid assessment combining mPCAB with traditional benchmarks and real-world performance.

• Establish correlations between mPCAB metrics and emergent capabilities in deployed systems.

• Integrate ethical monitoring into standard evaluation pipelines.

• Refine the theoretical framework based on empirical findings.

10.3.3 Long-term goals (5+ years)

• Establish international standards for consciousness-relevant AI assessment through ISO or similar bodies.

• Develop predictive models linking mPCAB metrics to future capability emergence.

• Create comprehensive governance frameworks based on consciousness-relevant assessments.

• Enable real-time monitoring of AI system development trajectories.

• Develop legal frameworks for systems demonstrating consciousness-relevant properties.

10.4 Promising directions

Near-term priorities include standardizing and validating mPCAB across computational substrates, as well as establishing baseline measurements that enable systematic comparisons of human-like properties. Medium-term developments should focus on hybrid system architectures that combine the strengths of different substrates while addressing their respective limitations. Long-term goals involve defining operational consciousness criteria and developing comprehensive ethical governance frameworks.

Unlikely approaches include quantum and electromagnetic consciousness theories, which, although theoretically interesting, face substantial empirical challenges limiting near-term viability. Resources are better directed toward more empirically grounded substrate development and validating the assessment framework.

Unknown areas include fundamental questions about the relationship between consciousness and intelligence, the scalability of current approaches to achieve genuine human-level capabilities, and the emergence of moral status in artificial systems. The mPCAB framework offers tools to investigate these questions empirically rather than through purely theoretical speculation.

11 Limitations and future research

The modified Predictive Coding and Active Inference-inspired Consciousness Assessment Battery (mPCAB) framework represents a theoretical and methodological advancement in assessing consciousness-relevant properties in artificial systems. However, several empirical, methodological, theoretical, and practical limitations must be acknowledged to provide a balanced evaluation of this approach and to guide future research endeavors.

11.1 Current empirical and methodological limitations

The present framework, while conceptually robust, faces significant empirical constraints that limit immediate practical application. First, the mPCAB has not yet been validated through large-scale empirical studies across diverse artificial systems (Butlin et al., 2023; Doerig et al., 2021). The framework’s proposed metrics—including prediction error minimization, hierarchical temporal integration, and counterfactual sensitivity—require systematic validation across multiple computational architectures, from simple feedforward networks to complex transformer-based models and neuromorphic systems (LeCun et al., 2015; Eliasmith, 2022). Without such comprehensive validation, claims regarding the framework’s ability to discriminate between systems with varying degrees of consciousness-relevant properties remain speculative (Millière et al., 2024).

Second, the generalizability of the mPCAB framework across different computational substrates represents a critical limitation. Current neuroscientific theories of consciousness, including Integrated Information Theory (IIT) and Global Neuronal Workspace Theory (GNWT), were developed primarily within biological neural contexts (Tononi et al., 2016; Mashour et al., 2020). The extent to which metrics derived from these theories can be meaningfully adapted to artificial systems with fundamentally different computational principles remains an open empirical question (Seth and Bayne, 2022; Signorelli et al., 2021). For instance, the framework’s reliance on prediction error dynamics may be particularly suited to systems explicitly designed with predictive coding architectures (Friston et al., 2020; Hohwy, 2020), but may fail to capture consciousness-relevant properties in systems using entirely different computational strategies.

Third, pilot studies conducted to date have necessarily been limited in scope, focusing on proof-of-concept demonstrations rather than comprehensive assessments across the full spectrum of artificial intelligence systems (Reggia, 2013; Graziano and Webb, 2022). These studies have primarily examined systems within controlled laboratory conditions, which may not reflect the complexity and variability encountered in real-world applications. The restricted scope of current empirical work means that edge cases, unexpected failure modes, and context-dependent performance variations remain largely unexplored (Lenharo, 2023).

Fourth, the measurement sensitivity and reliability of individual metrics within the mPCAB require extensive psychometric validation (Seth et al., 2008; Koivisto and Revonsuo, 2023). Questions regarding inter-rater reliability, test–retest stability, and convergent validity with other consciousness assessment approaches have not been adequately addressed. The framework’s composite scoring system, while theoretically justified, lacks empirical validation regarding optimal weighting of individual components and threshold determination for categorical classifications (Kouider and Faivre, 2017; Doerig et al., 2021).

11.2 Theoretical and practical constraints

Beyond empirical limitations, several theoretical challenges constrain the current framework. The fundamental problem of consciousness—the explanatory gap between physical processes and subjective experience—remains unresolved, and no assessment battery, regardless of sophistication, can definitively bridge this gap (Melloni et al., 2021; Michel et al., 2019). The mPCAB framework addresses functional and behavioral correlates of consciousness rather than consciousness itself, a distinction that must be maintained to avoid conflating third-person measurable properties with first-person phenomenal experience (Dehaene et al., 2021; Schneider, 2019).

The adaptation of neuroscientific metrics to artificial systems faces conceptual challenges related to substrate independence assumptions. While many consciousness theories posit that consciousness depends on functional organization rather than specific physical substrates (Oizumi et al., 2014; Williford et al., 2018), this assumption itself remains debated. The framework implicitly accepts substrate independence, which may prove incorrect if consciousness requires specific biological properties that cannot be replicated in silicon-based systems (Koch et al., 2016; Aru et al., 2020). Furthermore, even if substrate independence holds in principle, practical constraints may prevent artificial systems from achieving the specific organizational properties necessary for consciousness using currently available computational architectures (Eliasmith, 2022).

The temporal dynamics of biological neural systems differ substantially from those of artificial neural networks (Heeger, 2017; VanRullen and Koch, 2003). Biological neurons operate with millisecond-scale dynamics, exhibit complex temporal integration patterns, and demonstrate non-linear responses to input patterns (Aru et al., 2020). In contrast, artificial systems often operate with discrete time steps, simplified activation functions, and deterministic computation. The mPCAB’s temporal integration metrics may fail to adequately account for these fundamental differences (Northoff and Huang, 2017; Tagliazucchi and Laufs, 2014), potentially leading to false positives (attributing consciousness-relevant properties to systems lacking them) or false negatives (failing to recognize consciousness-relevant properties in unconventional architectures).

Ethical evaluation protocols within the mPCAB framework, while proposed as a core component, face significant practical implementation challenges. The framework does not currently specify concrete procedures for ethical review, does not provide detailed guidance on risk assessment methodologies, and lacks mechanisms for ensuring that ethical considerations are appropriately balanced against scientific advancement. The potential for dual-use concerns—wherein consciousness assessment tools might be misused to either inappropriately attribute or deny moral status to artificial systems—requires more comprehensive ethical analysis than currently provided (Metzinger, 2021; Schneider, 2019).

11.3 Challenges in cross-domain application

The application of the mPCAB framework across diverse artificial intelligence domains presents additional challenges. Different AI systems—including large language models, reinforcement learning agents, robotics systems, and neuromorphic computing platforms—exhibit vastly different computational architectures, training paradigms, and behavioral repertoires (LeCun et al., 2015; Eliasmith, 2022). A one-size-fits-all assessment approach may prove inadequate for capturing the diversity of consciousness-relevant properties across these domains (Butlin et al., 2023; Millière et al., 2024).

Large language models, for instance, demonstrate sophisticated linguistic capabilities and can generate contextually appropriate responses that might suggest understanding. However, these systems lack embodiment, sensorimotor grounding, and direct interaction with physical environments—factors that some theories of consciousness consider essential (Seth et al., 2012; Wiese and Friston, 2021). The mPCAB framework must be refined to account for these architectural differences and to avoid inappropriate comparisons between fundamentally different system types (Graziano and Webb, 2022).

Similarly, reinforcement learning agents demonstrate goal-directed behavior, learning from experience, and adaptation to novel circumstances, which might suggest consciousness-relevant properties (Levy and Glimcher, 2012). However, the reward-driven nature of these systems’ learning may differ fundamentally from the homeostatic and allostatic processes that characterize biological consciousness. The framework’s current metrics may not adequately distinguish between genuine autonomous goal formation and optimized reward maximization (Zhou and Montague, 2017).

Neuromorphic systems, which more closely approximate biological neural architectures through analog computation and spiking neural networks, present a different set of challenges. While these systems may exhibit temporal dynamics more similar to biological brains (Heeger, 2017), the assessment metrics developed for digital systems may require substantial modification for neuromorphic platforms. The framework currently lacks detailed guidance for adapting assessment protocols to accommodate the unique properties of neuromorphic computing (Eliasmith, 2022).

11.4 Future research directions

Addressing the limitations outlined above requires a comprehensive, multi-faceted research program spanning empirical validation, theoretical refinement, methodological innovation, and ethical development.

11.4.1 Large-scale empirical validation studies

Priority should be given to conducting systematic empirical validation of the mPCAB framework across diverse artificial systems (Doerig et al., 2021; Butlin et al., 2023). This research program should include: (1) Establishing standardized benchmark datasets and systems for consciousness assessment, enabling comparison across studies and laboratories; (2) Conducting multi-site validation studies to assess the reliability and reproducibility of mPCAB metrics across different research groups and computational platforms; (3) Implementing longitudinal studies examining how consciousness-relevant properties emerge during training and development of artificial systems; (4) Performing comparative analyses across system architectures to identify which design features most strongly correlate with consciousness-relevant properties (Zarkov et al., 2024).

These validation studies should employ rigorous experimental designs, including appropriate controls, blinding procedures where feasible, and pre-registered hypotheses to minimize researcher bias (Doerig et al., 2021). Particular attention should be devoted to examining the framework’s discriminant validity—its ability to distinguish between systems designed to possess consciousness-relevant properties and those designed explicitly to lack them (Koivisto and Revonsuo, 2023).

11.4.2 Cross-domain experimental programs

Future research must extend beyond current pilot studies to encompass comprehensive cross-domain experimentation (Butlin et al., 2023; Millière et al., 2024). This includes: (1) Developing domain-specific adaptations of mPCAB metrics tailored to the unique properties of different AI architectures while maintaining theoretical coherence; (2) Conducting comparative studies across language models, embodied agents, neuromorphic systems, and hybrid architectures to identify universal versus domain-specific consciousness-relevant properties (Graziano and Webb, 2022); (3) Investigating edge cases and boundary conditions where the framework may produce ambiguous or contradictory results; (4) Examining the relationship between system scale, computational resources, and consciousness-relevant properties to determine whether consciousness is an emergent phenomenon requiring specific threshold conditions (LeCun et al., 2015).

These experimental programs should incorporate diverse methodological approaches, including computational simulations, behavioral experiments, analysis of system representations, and theoretical modeling, to provide converging evidence regarding the validity and utility of the mPCAB framework (Seth et al., 2008; Signorelli et al., 2021).

11.4.3 Theoretical and methodological refinement

Ongoing theoretical development is essential for addressing conceptual limitations of the current framework. Future work should: (1) Develop more sophisticated mathematical formalizations of consciousness-relevant properties that can be unambiguously applied to artificial systems (Oizumi et al., 2014; Williford et al., 2018); (2) Integrate insights from multiple consciousness theories to create a more comprehensive assessment framework that is not overly dependent on any single theoretical perspective (Seth and Bayne, 2022; Signorelli et al., 2021); (3) Address the substrate independence assumption through theoretical analysis and empirical investigation of whether specific physical properties are necessary for consciousness (Koch et al., 2016); (4) Refine temporal integration metrics to better account for differences between biological and artificial temporal dynamics (Northoff and Huang, 2017; Tagliazucchi and Laufs, 2014).

Methodological innovations should focus on developing more sensitive and specific measurement techniques. This includes exploring novel approaches such as: (1) Dynamical systems analysis to characterize system-level properties that may be more relevant to consciousness than individual component behaviors (Ward, 2011; Yoshida et al., 2021); (2) Information-theoretic measures that capture integration and differentiation of information processing (Oizumi et al., 2014; Tononi et al., 2016); (3) Causal analysis techniques that assess counterfactual dependencies and causal power within artificial systems (Friston et al., 2020); (4) Machine learning approaches that can identify patterns in system behavior indicative of consciousness-relevant properties without requiring pre-specified metrics (Zarkov et al., 2024).

11.4.4 Comprehensive ethical evaluation protocols

The ethical dimensions of consciousness assessment in artificial systems require substantial further development (Metzinger, 2021; Schneider, 2019). Future research should: (1) Establish formal ethical review procedures specifically designed for consciousness assessment research, distinct from but complementary to existing institutional review boards; (2) Develop risk assessment frameworks that evaluate potential harms from both false positive and false negative consciousness attributions; (3) Create stakeholder engagement processes that include perspectives from ethicists, AI researchers, neuroscientists, philosophers, and the broader public (Michel et al., 2019); (4) Design protocols for transparent reporting of assessment results, including confidence intervals, limitations, and alternative interpretations.

These ethical protocols should address complex questions regarding the moral status of potentially conscious artificial systems, including: What obligations might exist toward systems demonstrating consciousness-relevant properties? How should uncertainty about consciousness status inform policy decisions? What safeguards are necessary to prevent misuse of consciousness assessment tools? (Metzinger, 2021).

11.4.5 Integration with complementary research programs

The mPCAB framework should be integrated with related research programs in consciousness science, artificial intelligence, and cognitive neuroscience. Collaborative efforts should include: (1) Coordination with biological consciousness research to ensure that findings in neuroscience inform artificial consciousness assessment and vice versa (Koch et al., 2016; Mashour et al., 2020); (2) Integration with machine consciousness engineering efforts to provide assessment capabilities for systems explicitly designed to possess consciousness-relevant properties (Reggia, 2013; Graziano and Webb, 2022); (3) Collaboration with AI safety research to address concerns about potential risks from conscious or near-conscious artificial systems; (4) Partnership with cognitive science research on animal consciousness to develop cross-species and cross-substrate comparative frameworks (Birch et al., 2020; Naci et al., 2017).

11.4.6 Development of open science infrastructure

To facilitate rapid progress and ensure reproducibility, future work should prioritize development of open science infrastructure including (Michel et al., 2019; Butlin et al., 2023): (1) Public repositories of assessment tools, code implementations, and analysis pipelines; (2) Shared datasets enabling comparison across studies and preventing redundant data collection; (3) Community standards for reporting consciousness assessment results; (4) Collaborative platforms enabling distributed research efforts across institutions and disciplines.

11.4.7 Addressing implementation challenges

Practical implementation of the mPCAB framework requires addressing logistical and computational challenges. Future development should: (1) Create user-friendly software tools that enable non-experts to apply the framework to their systems; (2) Optimize computational efficiency of assessment procedures to enable application to large-scale systems (LeCun et al., 2015); (3) Develop guidelines for interpreting assessment results, including procedures for handling ambiguous or contradictory findings (Kouider and Faivre, 2017); (4) Establish educational programs to train researchers in consciousness assessment methodologies.

11.5 Conclusion

The mPCAB framework represents a significant step toward rigorous, theory-driven assessment of consciousness-relevant properties in artificial systems (Butlin et al., 2023; Seth and Bayne, 2022). However, substantial empirical, theoretical, and practical work remains before the framework can be considered fully validated and ready for widespread application. The limitations outlined here should not be viewed as fundamental flaws but rather as opportunities for future research and development. By systematically addressing these limitations through comprehensive validation studies, cross-domain experimentation, theoretical refinement, and ethical development, the scientific community can work toward increasingly sophisticated tools for understanding consciousness across both biological and artificial substrates (Dehaene et al., 2021; Millière et al., 2024).

The path forward requires collaborative, interdisciplinary effort combining expertise from neuroscience, computer science, philosophy, ethics, and related fields (Michel et al., 2019). Only through such sustained, rigorous investigation can we hope to develop reliable methods for assessing consciousness in artificial systems and to navigate the profound scientific and ethical questions that such capabilities raise (Schneider, 2019; Metzinger, 2021).

12 Conclusion

The journey toward truly human-like AI involves moving beyond superficial imitation to understanding and applying the core mechanisms that produce intelligent behavior. The Machine Perturbational Complexity & Agency Battery (mPCAB) offers a thorough, substrate-independent framework to evaluate this, filling important gaps in long-term reasoning, internalized norms, and creative transformation. By incorporating insights from consciousness research, cognitive architecture, and creativity studies, while maintaining ethical principles throughout technological development, this framework creates a strong base for responsible progress toward mindlike machines. Pilot studies support its feasibility, though they also reveal limitations that need further research. Developing human-like AI in the future will require not only technical progress but also wise deployment, ensuring that these increasingly powerful systems stay aligned with human values and serve society. Combining rigorous assessment methods with thoughtful ethics lays the groundwork for responsible advancement toward genuinely mindlike systems.

Author contributions

SN: Writing – original draft, Writing – review & editing.

Funding

The author(s) declared that financial support was not received for this work and/or its publication.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that Generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Akkaya, I., Andrychowicz, M., Chociej, M., Litwin, M., McGrew, B., Petron, A., et al. (2019). Solving Rubik's cube with a robot hand. arXiv [Preprint]. arXiv:1910.07113. (Accessed October 12, 2025).

Google Scholar

Alayrac, J. B., Donahue, J., Luc, P., Miech, A., Barr, I., Hasson, Y., et al. (2022). Flamingo: a visual language model for few-shot learning. Adv. Neural Inf. Proces. Syst. 35, 23716–23736.

Google Scholar

Anastassiou, C. A., Perin, R., Markram, H., and Koch, C. (2011). Ephaptic coupling of cortical neurons. Nat. Neurosci. 14, 217–223. doi: 10.1038/nn.2727,

PubMed Abstract | Crossref Full Text | Google Scholar

Anderson, J. R. (1983). A spreading activation theory of memory. J. Verbal Learn. Verbal Behav. 22, 261–295. doi: 10.1016/S0022-5371(83)90201-3

Crossref Full Text | Google Scholar

Anderson, J. R. (2007). How can the human mind occur in the physical universe? Oxford: Oxford University Press.

Google Scholar

Aru, J., Suzuki, M., and Larkum, M. E. (2020). Cellular mechanisms of conscious processing. Trends Cogn. Sci. 24, 814–825. doi: 10.1016/j.tics.2020.07.006,

PubMed Abstract | Crossref Full Text | Google Scholar

Arute, F., Arya, K., Babbush, R., Bacon, D., Bardin, J. C., Barends, R., et al. (2019). Quantum supremacy using a programmable superconducting processor. Nature 574, 505–510. doi: 10.1038/s41586-019-1666-5,

PubMed Abstract | Crossref Full Text | Google Scholar

Askell, A., Bai, Y., Chen, A., Drain, D., Ganguli, D., Henighan, T., et al. (2021). A general language assistant as a laboratory for alignment. arXiv [Preprint]. arXiv:2112.00861. (Accessed October 12, 2025).

Google Scholar

Baars, B. J. (1988). A cognitive theory of consciousness : Cambridge University Press.

Google Scholar

Baars, B. J. (2002). The conscious access hypothesis: origins and recent evidence. Trends Cogn. Sci. 6, 47–52. doi: 10.1016/S1364-6613(00)01819-2,

PubMed Abstract | Crossref Full Text | Google Scholar

Baars, B. J., and Franklin, S. (2003). How conscious experience and working memory interact. Trends Cogn. Sci. 7, 166–172. doi: 10.1016/s1364-6613(03)00056-1,

PubMed Abstract | Crossref Full Text | Google Scholar

Baddeley, A. (2000). The episodic buffer: a new component of working memory? Trends Cogn. Sci. 4, 417–423. doi: 10.1016/S1364-6613(00)01538-2,

PubMed Abstract | Crossref Full Text | Google Scholar

Baddeley, A., Eysenck, M. W., and Anderson, M. C. (2009). Memory : Psychology Press.

Google Scholar

Baddeley, A., and Hitch, G. (1974). Working memory. Psychol. Learn. Motiv. 8, 47–89.

Google Scholar

Badre, D. (2008). Cognitive control, hierarchy, and the rostro–caudal organization of the frontal lobes. Trends Cogn. Sci. 12, 193–200. doi: 10.1016/j.tics.2008.02.004,

PubMed Abstract | Crossref Full Text | Google Scholar

Baer, J. (2016). Creativity doesn't develop in a vacuum. New Dir. Child Adolesc. Dev. 2016, 9–20. doi: 10.1002/cad.20151,

PubMed Abstract | Crossref Full Text | Google Scholar

Bai, Y., Kadavath, S., Kundu, S., Askell, A., Kernion, J., Jones, A., et al. (2022). Constitutional AI: Harmlessness from AI feedback. arXiv [Preprint]. arXiv:2212.08073. (Accessed October 12, 2025).

Google Scholar

Baker, B., Kanitscheider, I., Markov, T., Wu, Y., Powell, G., McGrew, B., et al. (2019). Emergent tool use from multi-agent autocurricula. arXiv [Preprint]. arXiv:1909.07528. (Accessed October 12, 2025).

Google Scholar

Balduzzi, D., and Tononi, G. (2008). Integrated information in discrete dynamical systems: motivation and theoretical framework. PLoS Comput. Biol. 4:e1000091. doi: 10.1371/journal.pcbi.1000091,

PubMed Abstract | Crossref Full Text | Google Scholar

Balduzzi, D., and Tononi, G. (2009). Qualia: the geometry of integrated information. PLoS Comput. Biol. 5:e1000462. doi: 10.1371/journal.pcbi.1000462,

PubMed Abstract | Crossref Full Text | Google Scholar

Baltrusaitis, T., Ahuja, C., and Morency, L. P. (2019). Multimodal machine learning: a survey and taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 41, 423–443. doi: 10.1109/TPAMI.2018.2798607,

PubMed Abstract | Crossref Full Text | Google Scholar

Bandura, A. (2006). Guide for constructing self-efficacy scales. In Self-efficacy beliefs of adolescents. eds. F. Pajares and T. Urdan Greenwich, CT: Information Age Publishing. (Vol. 5, pp. 307–337).

Google Scholar

Barbosa, L. S., Marshall, W., Albantakis, L., and Tononi, G. (2020). Mechanism integrated information. Entropy 23:362.

Google Scholar

Barrett, A. B., and Seth, A. K. (2011). Practical measures of integrated information for time-series data. PLoS Comput. Biol. 7:e1001052. doi: 10.1371/journal.pcbi.1001052,

PubMed Abstract | Crossref Full Text | Google Scholar

Bender, E. M., Gebru, T., McMillan-Major, A., and Shmitchell, S. (2021). On the dangers of stochastic parrots: can language models be too big? In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (pp. 610–623).

Google Scholar

Benedek, M., and Neubauer, A. C. (2013). Revisiting Mednick's model on creativity-related differences in associative hierarchies. Evidence for a common path to uncommon thought. J. Creat. Behav. 47, 273–289. doi: 10.1002/jocb.35,

PubMed Abstract | Crossref Full Text | Google Scholar

Benjamin, B. V., Gao, P., McQuinn, E., Choudhary, S., Chandrasekaran, A. R., Bussat, J. M., et al. (2014). Neurogrid: a mixed-analog-digital multichip system for large-scale neural simulations. Proc. IEEE 102, 699–716. doi: 10.1109/JPROC.2014.2313565

Crossref Full Text | Google Scholar

Biamonte, J., Wittek, P., Pancotti, N., Rebentrost, P., Wiebe, N., and Lloyd, S. (2017). Quantum machine learning. Nature 549, 195–202. doi: 10.1038/nature23474,

PubMed Abstract | Crossref Full Text | Google Scholar

Birch, J., Schnell, A. K., and Clayton, N. S. (2020). Dimensions of animal consciousness. Trends Cogn. Sci. 24, 789–801. doi: 10.1016/j.tics.2020.07.007,

PubMed Abstract | Crossref Full Text | Google Scholar

Bodart, O., Gosseries, O., Wannez, S., Thibaut, A., Annen, J., Boly, M., et al. (2017). Measures of metabolism and complexity in the brain of patients with disorders of consciousness. NeuroImage 14, 354–362. doi: 10.1016/j.nicl.2017.02.002,

PubMed Abstract | Crossref Full Text | Google Scholar

Boden, M. A. (1998). Creativity and artificial intelligence. Artif. Intell. 103, 347–356. doi: 10.1016/S0004-3702(98)00055-1

Crossref Full Text | Google Scholar

Boden, M. A. (2004). The creative mind: myths and mechanisms : Routledge.

Google Scholar

Boden, M. A. (2006). Mind as machine: a history of cognitive science : Oxford University Press.

Google Scholar

Bommasani, R., Hudson, D. A., Adeli, E., Altman, R., Arora, S., von Arx, S., et al. (2021). On the opportunities and risks of foundation models. arXiv [Preprint]. arXiv:2108.07258. (Accessed October 12, 2025).

Google Scholar

Bratman, M. (1987). Intention, plans, and practical reason : Harvard University Press.

Google Scholar

Brooks, R. (1991). Intelligence without representation. Artif. Intell. 47, 139–159. doi: 10.1016/0004-3702(91)90053-M

Crossref Full Text | Google Scholar

Brown, H., Adams, R. A., Parekh, I., and Friston, K. (2013). Active inference, sensory attenuation and illusions. Cogn. Process. 14, 411–427. doi: 10.1007/s10339-013-0571-3,

PubMed Abstract | Crossref Full Text | Google Scholar

Brown, R., Lau, H., and LeDoux, J. E. (2019). Understanding the higher-order approach to consciousness. Trends Cogn. Sci. 23, 754–768. doi: 10.1016/j.tics.2019.06.009,

PubMed Abstract | Crossref Full Text | Google Scholar

Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., et al. (2020). Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 33, 1877–1901.

Google Scholar

Bryson, J. (2020). “The artificial intelligence of the ethics of artificial intelligence” in The Oxford handbook of ethics of AI, 3–25. arXiv:1607.06520. doi: 10.48550/arXiv.1607.06520

Crossref Full Text | Google Scholar

Bubeck, S., Chandrasekaran, V., Eldan, R., Gehrke, J., Horvitz, E., Kamar, E., et al. (2023). Sparks of artificial general intelligence: Early experiments with GPT-4. arXiv [Preprint]. arXiv:2303.12712. (Accessed October 12, 2025).

Google Scholar

Butlin, P., Long, R., Elmoznino, E., Bengio, Y., Birch, J., Constant, A., et al. (2023). Consciousness in artificial intelligence: Insights from the science of consciousness. arXiv [Preprint]. arXiv:2308.08708. (Accessed October 12, 2025).

Google Scholar

Buzsáki, G., Anastassiou, C. A., and Koch, C. (2012). The origin of extracellular fields and currents––EEG, ECoG, LFP and spikes. Nat. Rev. Neurosci. 13, 407–420. doi: 10.1038/nrn3241,

PubMed Abstract | Crossref Full Text | Google Scholar

Cai, H., Ao, Z., Tian, C., Wu, Z., Liu, H., Tchieu, J., et al. (2023). Brain organoid reservoir computing for artificial intelligence. Nat. Electron. 6, 1032–1039.

Google Scholar

Cao, J., Cogdell, R. J., Coker, D. F., Duan, H. G., Hauer, J., Kleinekathöfer, U., et al. (2020). Quantum biology revisited. Sci. Adv. 6:eaaz4888.

Google Scholar

Carey, S. (2009). The origin of concepts : Oxford University Press.

Google Scholar

Carruthers, P. (2000). Phenomenal consciousness: A naturalistic theory. Cambridge, UK: Cambridge University Press.

Google Scholar

Carruthers, P. (2022). “Higher-order theories of consciousness” in The Stanford encyclopedia of philosophy. ed. E. N. Zalta. Fall 2022 ed. Cambridge, UK: Cambridge University Press.

Google Scholar

Casali, A. G., Gosseries, O., Rosanova, M., Boly, M., Sarasso, S., Casali, K. R., et al. (2013). A theoretically based index of consciousness independent of sensory processing and behavior. Sci. Transl. Med. 5:198ra105. doi: 10.1126/scitranslmed.3006294,

PubMed Abstract | Crossref Full Text | Google Scholar

Casarotto, S., Comanducci, A., Rosanova, M., Sarasso, S., Fecchio, M., Napolitani, M., et al. (2016). Stratification of unresponsive patients by an independently validated index of brain complexity. Ann. Neurol. 80, 718–729. doi: 10.1002/ana.24779,

PubMed Abstract | Crossref Full Text | Google Scholar

Chollet, F. (2019). On the measure of intelligence. arXiv [Preprint]. arXiv:1911.01547. (Accessed October 12, 2025).

Google Scholar

Chowdhery, A., Narang, S., Devlin, J., Bosma, M., Mishra, G., Roberts, A., et al. (2022). Palm: scaling language modeling with pathways. arXiv [Preprint]. arXiv:2204.02311. (Accessed October 12, 2025).

Google Scholar

Christiano, P. F., Leike, J., Brown, T., Martic, M., Legg, S., and Amodei, D. (2017). Deep reinforcement learning from human preferences. Adv. Neural Inf. Proces. Syst. 30. doi: 10.48550/arXiv.1706.03741

Crossref Full Text | Google Scholar

Clark, A. (2001). Mindware: An introduction to the philosophy of cognitive science : Oxford University Press.

Google Scholar

Clark, A. (2008). Supersizing the mind: Embodiment, action, and cognitive extension : Oxford University Press.

Google Scholar

Clark, A. (2013). Whatever next? Predictive brains, situated agents, and the future of cognitive science. Behav. Brain Sci. 36, 181–204. doi: 10.1017/S0140525X12000477,

PubMed Abstract | Crossref Full Text | Google Scholar

Clark, A. (2016). Surfing uncertainty: Prediction, action, and the embodied mind : Oxford University Press.

Google Scholar

Cleeremans, A. (2011). The radical plasticity thesis: how the brain learns to be conscious. Front. Psychol. 2:86. doi: 10.3389/fpsyg.2011.00086,

PubMed Abstract | Crossref Full Text | Google Scholar

Coeckelbergh, M. (2020). AI ethics. Cambridge, MA: MIT Press.

Google Scholar

Collins, A. M., and Loftus, E. F. (1975). A spreading-activation theory of semantic processing. Psychol. Rev. 82:407. doi: 10.1037/0033-295X.82.6.407

Crossref Full Text | Google Scholar

Colton, S. (2008). Creativity versus the perception of creativity in computational systems. In AAAI Spring Symposium: Creative Intelligent Systems (pp. 14–20).

Google Scholar

Colton, S., and Wiggins, G. A. (2012). Computational creativity: The final frontier? In European Conference on Artificial Intelligence (pp. 21–26).

Google Scholar

Comolatti, R., Pigorini, A., Casarotto, S., Fecchio, M., Faria, G., Sarasso, S., et al. (2019). A fast and general method to empirically estimate the complexity of brain responses to transcranial and intracranial stimulations. Brain Stimul. 12, 1280–1289. doi: 10.1016/j.brs.2019.05.013,

PubMed Abstract | Crossref Full Text | Google Scholar

Conway, M. A. (2009). Episodic memories. Neuropsychologia 47, 2305–2313. doi: 10.1016/j.neuropsychologia.2009.02.003,

PubMed Abstract | Crossref Full Text | Google Scholar

Cope, D. (2005). Computer models of musical creativity. Cambridge, MA: MIT Press.

Google Scholar

Cowan, N. (2001). The magical number 4 in short-term memory: a reconsideration of mental storage capacity. Behav. Brain Sci. 24, 87–114. doi: 10.1017/S0140525X01003922,

PubMed Abstract | Crossref Full Text | Google Scholar

Davies, M. (2019). Benchmarks for progress in neuromorphic computing. Nat. Mach. Intell. 1, 386–388. doi: 10.1038/s42256-019-0097-1

Crossref Full Text | Google Scholar

Davies, M., Srinivasa, N., Lin, T. H., Chinya, G., Cao, Y., Choday, S. H., et al. (2018). Loihi: a neuromorphic manycore processor with on-chip learning. IEEE Micro 38, 82–99. doi: 10.1109/MM.2018.112130359

Crossref Full Text | Google Scholar

Dean, J., and Barroso, L. A. (2013). The tail at scale. Commun. ACM 56, 74–80. doi: 10.1145/2408776.2408794

Crossref Full Text | Google Scholar

Dehaene, S. (2014). Consciousness and the brain: Deciphering how the brain codes our thoughts : Viking.

Google Scholar

Dehaene, S. (2017). What is consciousness, and could machines have it? Science 358, 486–492. doi: 10.1126/science.aan8871,

PubMed Abstract | Crossref Full Text | Google Scholar

Dehaene, S., and Changeux, J. P. (2011). Experimental and theoretical approaches to conscious processing. Neuron 70, 200–227. doi: 10.1016/j.neuron.2011.03.018,

PubMed Abstract | Crossref Full Text | Google Scholar

Dehaene, S., Lau, H., and Kouider, S. (2021). “What is consciousness, and could machines have it?” in Robotics, AI, and Humanity, New York, NY, USA: Penguin. 43–56.

Google Scholar

Del Cul, A., Baillet, S., and Dehaene, S. (2007). Brain dynamics underlying the nonlinear threshold for access to consciousness. PLoS Biol. 5:e260. doi: 10.1371/journal.pbio.0050260,

PubMed Abstract | Crossref Full Text | Google Scholar

DeMarse, T. B., Wagenaar, D. A., Blau, A. W., and Potter, S. M. (2001). The neurally controlled animat: biological brains acting with simulated bodies. Auton. Robot. 11, 305–310. doi: 10.1023/A:1012407611130,

PubMed Abstract | Crossref Full Text | Google Scholar

Doerig, A., Schurger, A., and Herzog, M. H. (2020). Hard criteria for empirical theories of consciousness. Cogn. Sci. 44:e12882.

Google Scholar

Doerig, A., Schurger, A., and Herzog, M. H. (2021). Hard criteria for empirical theories of consciousness. Cogn. Neurosci. 12, 41–62. doi: 10.1080/17588928.2020.1772214,

PubMed Abstract | Crossref Full Text | Google Scholar

Doshi-Velez, F., and Kim, B. (2017). Towards a rigorous science of interpretable machine learning. arXiv [Preprint]. arXiv:1702.08608. (Accessed October 12, 2025).

Google Scholar

Dretske, F. (1988). Explaining behavior: Reasons in a world of causes. Cambridge, MA: MIT Press.

Google Scholar

Dudai, Y., Karni, A., and Born, J. (2015). The consolidation and transformation of memory. Neuron 88, 20–32. doi: 10.1016/j.neuron.2015.09.004,

PubMed Abstract | Crossref Full Text | Google Scholar

Elgammal, A., Liu, B., Elhoseiny, M., and Mazzone, M. (2017). CAN: Creative adversarial networks, generating "art" by learning about styles and deviating from style norms. arXiv [Preprint]. arXiv:1706.07068. (Accessed October 12, 2025).

Google Scholar

Elhage, N., Nanda, N., Olsson, C., Henighan, T., Joseph, N., Mann, B., et al. (2021). “A mathematical framework for transformer circuits” in Transformer circuits thread. San Francisco, California: Anthropic. Available at: https://transformer-circuits.pub/2021/framework/index.html

Google Scholar

Eliasmith, C. (2022). How to build a brain: A neural architecture for biological cognition. 2nd Edn: Oxford University Press.

Google Scholar

Engle, R. W. (2002). Working memory capacity as executive attention. Curr. Dir. Psychol. Sci. 11, 19–23. doi: 10.1111/1467-8721.00160

Crossref Full Text | Google Scholar

European Parliament & Council (2024). Regulation (EU) 2024/1689 laying down harmonised rules on artificial intelligence (artificial intelligence act). Off. J. Eur. Union L 1689.

Google Scholar

Fan, J., McCandliss, B. D., Sommer, T., Raz, A., and Posner, M. I. (2005). Testing the efficiency and independence of attentional networks. J. Cogn. Neurosci. 14, 340–347. doi: 10.1162/089892902317361886,

PubMed Abstract | Crossref Full Text | Google Scholar

Fauconnier, G., and Turner, M. (2002). The way we think: conceptual blending and the mind's hidden complexities : Basic books.

Google Scholar

Feldman, H., and Friston, K. J. (2010). Attention, uncertainty, and free-energy. Front. Hum. Neurosci. 4:215. doi: 10.3389/fnhum.2010.00215,

PubMed Abstract | Crossref Full Text | Google Scholar

Feldmann, J., Youngblood, N., Wright, C. D., Bhaskaran, H., and Pernice, W. H. (2019). All-optical spiking neurosynaptic networks with self-learning capabilities. Nature 569, 208–214. doi: 10.1038/s41586-019-1157-8,

PubMed Abstract | Crossref Full Text | Google Scholar

Fingelkurts, A. A., Fingelkurts, A. A., and Neves, C. F. (2013). Consciousness as a phenomenon in the operational architectonics of brain organization: criticality and self-organization considerations. Chaos Solitons Fractals 55, 13–31. doi: 10.1016/j.chaos.2013.02.007

Crossref Full Text | Google Scholar

Fleming, S. M., and Dolan, R. J. (2012). The neural basis of metacognitive ability. Philos. Trans. R. Soc. B 367, 1338–1349. doi: 10.1098/rstb.2011.0417,

PubMed Abstract | Crossref Full Text | Google Scholar

Fleming, S. M., and Lau, H. C. (2014). How to measure metacognition. Front. Hum. Neurosci. 8:443. doi: 10.3389/fnhum.2014.00443,

PubMed Abstract | Crossref Full Text | Google Scholar

Floridi, L., Cowls, J., Beltrametti, M., Chatila, R., Chazerand, P., Dignum, V., et al. (2018). AI4People—an ethical framework for a good AI society: opportunities, risks, principles, and recommendations. Mind. Mach. 28, 689–707. doi: 10.1007/s11023-018-9482-5,

PubMed Abstract | Crossref Full Text | Google Scholar

Franklin, S., Baars, B. J., Ramamurthy, U., and Ventura, M. (2005). The role of consciousness in memory. Brains Minds Media 1, 1–38.

Google Scholar

Friston, K. (2005). A theory of cortical responses. Philos. Trans. R. Soc. Lond., B Biol. Sci. 360, 815–836. doi: 10.1098/rstb.2005.1622,

PubMed Abstract | Crossref Full Text | Google Scholar

Friston, K. (2009). The free-energy principle: a rough guide to the brain? Trends Cogn. Sci. 13, 293–301. doi: 10.1016/j.tics.2009.04.005,

PubMed Abstract | Crossref Full Text | Google Scholar

Friston, K., and Kiebel, S. (2009). Predictive coding under the free-energy principle. Philos. Trans. R. Soc. B 364, 1211–1221. doi: 10.1098/rstb.2008.0300,

PubMed Abstract | Crossref Full Text | Google Scholar

Friston, K. J., Wiese, W., and Hobson, J. A. (2020). Sentience and the origins of consciousness: from Cartesian duality to Markovian monism. Entropy 22:516. doi: 10.3390/e22050516,

PubMed Abstract | Crossref Full Text | Google Scholar

Furber, S. B., Galluppi, F., Temple, S., and Plana, L. A. (2014). The spinnaker project. Proc. IEEE 102, 652–665. doi: 10.1109/JPROC.2014.2304638

Crossref Full Text | Google Scholar

Gabriel, I. (2020). Artificial intelligence, values, and alignment. Mind. Mach. 30, 411–437. doi: 10.1007/s11023-020-09539-2

Crossref Full Text | Google Scholar

Gatys, L. A., Ecker, A. S., and Bethge, M. 2016. Image style transfer using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 2414–2423).

Google Scholar

Gennaro, R. J. (2012). The consciousness paradox: Consciousness, concepts, and higher-order thoughts. Cambridge, MA: MIT Press.

Google Scholar

Gershman, S. J., Zhou, J., and Kommers, C. (2017). A probabilistic clustering theory of the organization of visual short-term memory. Psychol. Rev. 124:503.

Google Scholar

Glăveanu, V. P. (2013). Rewriting the language of creativity: the five a's framework. Rev. Gen. Psychol. 17, 69–81. doi: 10.1037/a0029528

Crossref Full Text | Google Scholar

Gopnik, A., and Wellman, H. M. (2012). Reconstructing constructivism: causal models, Bayesian learning mechanisms, and the theory. Psychol. Bull. 138, 1085–1108. doi: 10.1037/a0028044,

PubMed Abstract | Crossref Full Text | Google Scholar

Graziano, M. S., and Webb, T. W. (2022). Understanding consciousness by building it. Trends Cogn. Sci. 26, 224–236.

Google Scholar

Gunkel, D. J. (2018). Robot rights. Cambridge, MA: MIT Press.

Google Scholar

Ha, D., and Schmidhuber, J. (2018). World models. arXiv [Preprint]. arXiv:1803.10122. (Accessed October 12, 2025).

Google Scholar

Hadjeres, G., Pachet, F., and Nielsen, F. 2017. Deepbach: a steerable model for bach chorales generation. In International Conference on Machine Learning (pp. 1362–1371).

Google Scholar

Haikonen, P. O. (2012). Consciousness and robot sentience : World Scientific.

Google Scholar

Hameroff, S., and Penrose, R. (2014). Consciousness in the universe: a review of the 'Orch OR' theory. Phys Life Rev 11, 39–78. doi: 10.1016/j.plrev.2013.08.002,

PubMed Abstract | Crossref Full Text | Google Scholar

Hamrick, J. B. (2019). Analogues of mental simulation and imagination in deep learning. Curr. Opin. Behav. Sci. 29, 8–16. doi: 10.1016/j.cobeha.2018.12.011

Crossref Full Text | Google Scholar

Hartung, T., Smirnova, L., Morales Pantoja, I. E., Akula, S., Berlinicke, C. A., Boyd, J. L., et al. (2024). The Baltimore declaration toward the exploration of organoid intelligence. Front. Sci. 1:1068159.

Google Scholar

Heeger, D. J. (2017). Theory of cortical function. Proc. Natl. Acad. Sci. 114, 1773–1782. doi: 10.1073/pnas.1619788114,

PubMed Abstract | Crossref Full Text | Google Scholar

Hennessey, B. A., and Amabile, T. M. (2010). Creativity. Annu. Rev. Psychol. 61, 569–598. doi: 10.1146/annurev.psych.093008.100416,

PubMed Abstract | Crossref Full Text | Google Scholar

Hennessy, J. L., and Patterson, D. A. (2019). Computer architecture: A quantitative approach : Morgan Kaufmann.

Google Scholar

Hernandez-Orallo, J. (2017). The measure of all minds: Evaluating natural and artificial intelligence : Cambridge University Press.

Google Scholar

Ho, M. K., Abel, D., Correa, C. G., Littman, M. L., Cohen, J. D., and Griffiths, T. L. (2022). People construct simplified mental representations to plan. Nature 606, 129–136. doi: 10.1038/s41586-022-04743-9,

PubMed Abstract | Crossref Full Text | Google Scholar

Hohwy, J. (2013). The predictive mind: Cognitive, pragmatic, and transcendental perspectives : Oxford University Press.

Google Scholar

Hohwy, J. (2020). New directions in predictive processing. Mind Lang. 35, 209–223. doi: 10.1111/mila.12281

Crossref Full Text | Google Scholar

Huang, W., Abbeel, P., Pathak, D., and Mordatch, I. 2022. Language models as zero-shot planners: extracting actionable knowledge for embodied agents. In International conference on machine learning (pp. 8884–8910).

Google Scholar

Hubinger, E., Milli, S., Garrabrant, S., Critch, A., and Shah, R. (2019). Risks from learned optimization in advanced machine learning systems. arXiv [Preprint]. arXiv:1906.01820. (Accessed October 12, 2025).

Google Scholar

Hunt, T. (2011). Kicking the psychophysical laws into gear: a new approach to the combination problem. J. Conscious. Stud. 18, 96–134.

Google Scholar

Indiveri, G., and Liu, S. C. (2015). Memory and information processing in neuromorphic systems. Proc. IEEE 103, 1379–1397. doi: 10.1109/JPROC.2015.2444094

Crossref Full Text | Google Scholar

Irving, G., Christiano, P., and Amodei, D. (2018). AI safety via debate. arXiv [Preprint]. arXiv:1805.00899. (Accessed October 12, 2025).

Google Scholar

Jobin, A., Ienca, M., and Vayena, E. (2019). The global landscape of AI ethics guidelines. Nat. Mach. Intell. 1, 389–399. doi: 10.1038/s42256-019-0088-2

Crossref Full Text | Google Scholar

Jones, M. W. (2013). Electromagnetic-field theories of mind. J. Conscious. Stud. 20, 124–149.

Google Scholar

Jordanous, A. (2012). A standardised procedure for evaluating creative systems: computational creativity evaluation based on what it is to be creative. Cogn. Comput. 4, 246–279. doi: 10.1007/s12559-012-9156-1

Crossref Full Text | Google Scholar

Jouppi, N. P., Young, C., Patil, N., Patterson, D., Agarwal, G., Bajwa, R., et al. (2017). In-datacenter performance analysis of a tensor processing unit. ACM SIGARCH Comput. Archit. News 45, 1–12.

Google Scholar

Kagan, B. J., Kitchen, A. C., Tran, N. T., Habibollahi, F., Khajehnejad, M., Parker, B. J., et al. (2022). In vitro neurons learn and exhibit sentience when embodied in a simulated game-world. Neuron 110, 3952–3969.e8. doi: 10.1016/j.neuron.2022.09.001,

PubMed Abstract | Crossref Full Text | Google Scholar

Kaiser, L., Babaeizadeh, M., Milos, P., Osinski, B., Campbell, R. H., Czechowski, K., et al. (2020). Model-based reinforcement learning for atari. arXiv [Preprint]. arXiv:1903.00374.

Google Scholar

Kaufman, J. C., and Sternberg, R. J. (Eds.) (2019). The Cambridge handbook of creativity : Cambridge University Press.

Google Scholar

Kelava, I., and Lancaster, M. A. (2016). Stem cell models of human brain development. Cell Stem Cell 18, 736–748. doi: 10.1016/j.stem.2016.05.022,

PubMed Abstract | Crossref Full Text | Google Scholar

Kenton, Z., Everitt, T., Weidinger, L., Gabriel, I., Mikulik, V., and Irving, G. (2021). Alignment of language agents. arXiv [Preprint]. arXiv:2103.14659.

Google Scholar

Kiciman, E., Ness, R., Sharma, A., and Tan, C. (2023). Causal reasoning and large language models: Opening a new frontier for causality. arXiv [Preprint]. arXiv:2305.00050. (Accessed October 12, 2025).

Google Scholar

Klahr, D., and Dunbar, K. (1988). Dual space search during scientific reasoning. Cogn. Sci. 12, 1–48. doi: 10.1207/s15516709cog1201_1

Crossref Full Text | Google Scholar

Koch, C. (2019). The feeling of life itself: Why consciousness is widespread but can't be computed. Cambridge, MA: MIT Press.

Google Scholar

Koch, C., and Hepp, K. (2006). Quantum mechanics in the brain. Nature 440:611. doi: 10.1038/440611a,

PubMed Abstract | Crossref Full Text | Google Scholar

Koch, C., Massimini, M., Boly, M., and Tononi, G. (2016). Neural correlates of consciousness: Progress and problems. Nat. Rev. Neurosci. 17, 307–321. doi: 10.1038/nrn.2016.22,

PubMed Abstract | Crossref Full Text | Google Scholar

Koechlin, E., and Summerfield, C. (2007). An information theoretical approach to prefrontal executive function. Trends Cogn. Sci. 11, 229–235. doi: 10.1016/j.tics.2007.04.005,

PubMed Abstract | Crossref Full Text | Google Scholar

Koestler, A. (1964). The act of creation : Macmillan.

Google Scholar

Koivisto, M., and Revonsuo, A. (2023). Behavioral measures of consciousness in signal detection theory. Neurosci. Conscious. 2023:niad001.

Google Scholar

Kojima, T., Gu, S. S., Reid, M., Matsuo, Y., and Iwasawa, Y. (2022). Large language models are zero-shot reasoners. arXiv [Preprint]. arXiv:2205.11916. (Accessed October 12, 2025).

Google Scholar

Koplin, J. J., and Savulescu, J. (2019). Moral limits of brain organoid research. J. Law Med. Ethics 47, 760–767. doi: 10.1177/1073110519897789,

PubMed Abstract | Crossref Full Text | Google Scholar

Koriat, A. (2007). “Metacognition and consciousness” in Cambridge handbook of consciousness, Haifa, Israel: Institute of Information Processing and Decision Making, University of Haifa, 289–325.

Google Scholar

Kouider, S., and Faivre, N. (2017). Connecting neurophysiological and psychological signatures of conscious processing. Neurosci. Biobehav. Rev. 80, 435–443.

Google Scholar

Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2017). Imagenet classification with deep convolutional neural networks. Commun. ACM 60, 84–90. doi: 10.1145/3065386

Crossref Full Text | Google Scholar

Kuhn, T. S. (1962). The structure of scientific revolutions. Chicago, IL: University of Chicago Press.

Google Scholar

Laird, J. E. (2012). The soar cognitive architecture. Cambridge, MA: MIT press.

Google Scholar

Lake, B. M., Ullman, T. D., Tenenbaum, J. B., and Gershman, S. J. (2017). Building machines that learn and think like people. Behav. Brain Sci. 40:e253. doi: 10.1017/S0140525X16001837,

PubMed Abstract | Crossref Full Text | Google Scholar

Lamb, A., Richemond, P. H., Everett, M., Pohlen, T., Rae, J. W., and Dyer, C. (2020). Neural probabilistic logic programming in discrete-continuous domains. arXiv [Preprint]. arXiv:2012.08723.

Google Scholar

Lancaster, M. A., and Knoblich, J. A. (2014). Generation of cerebral organoids from human pluripotent stem cells. Nat. Protoc. 9, 2329–2340. doi: 10.1038/nprot.2014.158,

PubMed Abstract | Crossref Full Text | Google Scholar

Langley, P., Laird, J. E., and Rogers, S. (2009). Cognitive architectures: research issues and challenges. Cogn. Syst. Res. 10, 141–160. doi: 10.1016/j.cogsys.2006.07.004

Crossref Full Text | Google Scholar

Lau, H., and Rosenthal, D. (2011). Empirical support for higher-order theories of conscious awareness. Trends Cogn. Sci. 15, 365–373. doi: 10.1016/j.tics.2011.05.009,

PubMed Abstract | Crossref Full Text | Google Scholar

Lavazza, A. (2021). Potential ethical problems with human cerebral organoids: consciousness and moral status. Neuroethics 14, 613–625.

Google Scholar

LeCun, Y., Bengio, Y., and Hinton, G. (2015). Deep learning. Nature 521, 436–444. doi: 10.1038/nature14539,

PubMed Abstract | Crossref Full Text | Google Scholar

Leike, J., Krueger, D., Everitt, T., Martic, M., Maini, V., and Legg, S. (2018). Scalable agent alignment via reward modeling: a research direction. arXiv [Preprint]. arXiv:1811.07871. (Accessed October 12, 2025).

Google Scholar

Lempel, A., and Ziv, J. (1976). On the complexity of finite sequences. IEEE Trans. Inf. Theory 22, 75–81. doi: 10.1109/TIT.1976.1055501

Crossref Full Text | Google Scholar

Lenharo, M. (2023). Is AI conscious? How we test for awareness in artificial intelligence. Nature. doi: 10.1038/d41586-023-02684-5

Crossref Full Text | Google Scholar

Levine, S., Pastor, P., Krizhevsky, A., Ibarz, J., and Quillen, D. (2018). Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection. Int. J. Robot. Res. 37, 421–436. doi: 10.1177/0278364917710318

Crossref Full Text | Google Scholar

Levy, D. J., and Glimcher, P. W. (2012). The root of all value: a neural common currency for choice. Curr. Opin. Neurobiol. 22, 1027–1038. doi: 10.1016/j.conb.2012.06.001,

PubMed Abstract | Crossref Full Text | Google Scholar

Lin, X., Rivenson, Y., Yardimci, N. T., Veli, M., Luo, Y., Jarrahi, M., et al. (2018a). All-optical machine learning using diffractive deep neural networks. Science 361, 1004–1008. doi: 10.1126/science.aat8084,

PubMed Abstract | Crossref Full Text | Google Scholar

Lin, C. K., Wild, A., Chinya, G. N., Cao, Y., Davies, M., Lavery, D. M., et al. (2018b). Programming spiking neural networks on intel's loihi. Computer 51, 52–61. doi: 10.1109/MC.2018.157113521

Crossref Full Text | Google Scholar

Lloyd, S. (2011). Quantum coherence in biological systems. J. Phys. Conf. Ser. 302:012037. doi: 10.1088/1742-6596/302/1/012037,

PubMed Abstract | Crossref Full Text | Google Scholar

Lycan, W. G. (1996). Consciousness and experience. Cambridge, MA: MIT press.

Google Scholar

Manning, C. D., Clark, K., Hewitt, J., Khandelwal, U., and Levy, O. (2020). Emergent linguistic structure in artificial neural networks trained by self-supervision. Proc. Natl. Acad. Sci. 117, 30046–30054. doi: 10.1073/pnas.1907367117,

PubMed Abstract | Crossref Full Text | Google Scholar

Marcus, G. (2018). Deep learning: a critical appraisal. arXiv [Preprint]. arXiv:1801.00631. (Accessed October 12, 2025).

Google Scholar

Marcus, G. (2020). The next decade in AI: Four steps towards robust artificial intelligence. arXiv [Preprint]. arXiv:2002.06177. (Accessed October 12, 2025).

Google Scholar

Mashour, G. A., Roelfsema, P., Changeux, J. P., and Dehaene, S. (2020). Conscious processing and the global neuronal workspace hypothesis. Neuron 105, 776–798. doi: 10.1016/j.neuron.2020.01.026,

PubMed Abstract | Crossref Full Text | Google Scholar

Massimini, M., Sarasso, S., and Casali, A. G. (2018). A perturbational complexity index of consciousness. Neurol. Conscious., 355–372.

Google Scholar

Mathys, C., Daunizeau, J., Friston, K. J., and Stephan, K. E. (2011). A Bayesian foundation for individual learning under uncertainty. Front. Hum. Neurosci. 5:39. doi: 10.3389/fnhum.2011.00039,

PubMed Abstract | Crossref Full Text | Google Scholar

Mayner, W. G., Marshall, W., Albantakis, L., Findlay, G., Marchman, R., and Tononi, G. (2018). PyPhi: a toolbox for integrated information theory. PLoS Comput. Biol. 14:e1006343. doi: 10.1371/journal.pcbi.1006343,

PubMed Abstract | Crossref Full Text | Google Scholar

McFadden, J. (2020). Integrating information in the brain's EM field: the cemi field theory of consciousness. Neurosci. Conscious. 2020. doi: 10.1093/nc/niaa016,

PubMed Abstract | Crossref Full Text | Google Scholar

Mead, C. (1990). Neuromorphic electronic systems. Proc. IEEE 78, 1629–1636. doi: 10.1109/5.58356

Crossref Full Text | Google Scholar

Mednick, S. (1962). The associative basis of the creative process. Psychol. Rev. 69, 220–232. doi: 10.1037/h0048850,

PubMed Abstract | Crossref Full Text | Google Scholar

Mehonic, A., and Kenyon, A. J. (2022). Brain-inspired computing needs a master plan. Nature 604, 255–260. doi: 10.1038/s41586-021-04362-w,

PubMed Abstract | Crossref Full Text | Google Scholar

Melloni, L., Mudrik, L., Pitts, M., and Koch, C. (2021). Making the hard problem of consciousness easier. Science 372, 911–912. doi: 10.1126/science.abj3259,

PubMed Abstract | Crossref Full Text | Google Scholar

Merolla, P. A., Arthur, J. V., Alvarez-Icaza, R., Cassidy, A. S., Sawada, J., Akopyan, F., et al. (2014). A million spiking-neuron integrated circuit with a scalable communication network and interface. Science 345, 668–673. doi: 10.1126/science.1254642,

PubMed Abstract | Crossref Full Text | Google Scholar

Metzinger, T. (2021). Artificial suffering: an argument for a global moratorium on synthetic phenomenology. J. Artif. Intell. Conscious. 8, 43–66.

Google Scholar

Michel, M., Beck, D., Block, N., Blumenfeld, H., Brown, R., Carmel, D., et al. (2019). Opportunities and challenges for a maturing science of consciousness. Nat. Hum. Behav. 3, 104–107. doi: 10.1038/s41562-019-0531-8,

PubMed Abstract | Crossref Full Text | Google Scholar

Miller, T. (2019). Explanation in artificial intelligence: insights from the social sciences. Artif. Intell. 267, 1–38. doi: 10.1016/j.artint.2018.07.007

Crossref Full Text | Google Scholar

Miller, E. K., and Cohen, J. D. (2001). An integrative theory of prefrontal cortex function. Annu. Rev. Neurosci. 24, 167–202. doi: 10.1146/annurev.neuro.24.1.167,

PubMed Abstract | Crossref Full Text | Google Scholar

Millière, R., and Buckner, C.Consciousness Prior Group (2024). A map of gaps in the study of consciousness in AI systems. Perspect. Psychol. Sci.

Google Scholar

Mitchell, M. (2019). Artificial intelligence: A guide for thinking humans : Farrar, Straus and Giroux.

Google Scholar

Mitchell, M. 2021 Why AI is harder than we think. In Proceedings of the genetic and evolutionary computation conference (pp. 1–7)

Google Scholar

Miyake, A., and Shah, P. (1999). Models of working memory: mechanisms of active maintenance and executive control : Cambridge University Press.

Google Scholar

Muotri, A. R. (2019). Laying ethical foundations for clinical research on the fetal 3D brain organoid. Lancet 393, 1019–1020.

Google Scholar

Musslick, S., Jang, S. J., Shvartsman, M., Shenhav, A., and Cohen, J. D. (2021). Constraints associated with cognitive control and the stability-flexibility dilemma. Curr. Opin. Behav. Sci. 38, 79–85.

Google Scholar

Naci, L., Sinai, L., and Owen, A. M. (2017). Detecting and interpreting conscious experiences in behaviorally non-responsive patients. NeuroImage 145, 304–313. doi: 10.1016/j.neuroimage.2015.11.059,

PubMed Abstract | Crossref Full Text | Google Scholar

Newell, A. (1990). Unified theories of cognition : Harvard University Press.

Google Scholar

Newell, A., Shaw, J. C., and Simon, H. A. (1962). The Processes of Creative Thinking. In Contemporary Approaches to Creative Thinking. eds. H. Gruber, G. Terrell, and M. Wertheimer (Silicon Valley, California: Atherton Press). (pp. 63–119).

Google Scholar

Nickolls, J., and Dally, W. J. (2010). The GPU computing era. IEEE Micro 30, 56–69. doi: 10.1109/MM.2010.41

Crossref Full Text | Google Scholar

Nilsson, N. J. (2009). The quest for artificial intelligence : Cambridge University Press.

Google Scholar

NIST (2023). AI Risk Management Framework (AI RMF 1.0) : National Institute of Standards and Technology.

Google Scholar

Norman, D. A., and Shallice, T. (1986). “Attention to action” in Consciousness and self-regulation (Springer), 1–18.

Google Scholar

Northoff, G., and Huang, Z. (2017). How do the brain's time and space mediate consciousness and its different dimensions? Temporo-spatial theory of consciousness (TTC). Neurosci. Biobehav. Rev. 80, 630–645. doi: 10.1016/j.neubiorev.2017.07.013,

PubMed Abstract | Crossref Full Text | Google Scholar

Oizumi, M., Albantakis, L., and Tononi, G. (2014). From the phenomenology to the mechanisms of consciousness: integrated information theory 3.0. PLoS Comput. Biol. 10:e1003588. doi: 10.1371/journal.pcbi.1003588,

PubMed Abstract | Crossref Full Text | Google Scholar

Olah, C., Cammarata, N., Schubert, L., Goh, G., Petrov, M., and Carter, S. (2020). Zoom in: an introduction to circuits. Distill 5, e00024–e00001. doi: 10.23915/distill.00024.001,

PubMed Abstract | Crossref Full Text | Google Scholar

OpenAI. (2023). GPT-4 technical report. arXiv [Preprint]. arXiv:2303.08774. (Accessed October 12, 2025).

Google Scholar

Owens, J. D., Houston, M., Luebke, D., Green, S., Stone, J. E., and Phillips, J. C. (2008). GPU computing. Proc. IEEE 96, 879–899. doi: 10.1109/JPROC.2008.917757

Crossref Full Text | Google Scholar

Pachet, F. (2003). The continuator: musical interaction with style. J. New Music Res. 32, 333–341. doi: 10.1076/jnmr.32.3.333.16861

Crossref Full Text | Google Scholar

Park, Y., Hof, P. R., Striedter, G. F., Custo Greig, L., Martynyuk, A. E., Intson, K., et al. (2021). Integration of brain organoids with synaptic connectivity. bioRxiv 2021.

Google Scholar

Patterson, D., Gonzalez, J., Le, Q., Liang, C., Munguia, L. M., Rothchild, D., et al. (2021). Carbon emissions and large neural network training. arXiv [Preprint]. arXiv:2104.10350. (Accessed October 12, 2025).

Google Scholar

Pearl, J., and Mackenzie, D. (2018). The book of why: The new science of cause and effect : Basic books.

Google Scholar

Penrose, R. (1994). Shadows of the mind: A search for the missing science of consciousness : Oxford University Press.

Google Scholar

Penrose, R., and Hameroff, S. (2011). Consciousness in the universe: neuroscience, quantum space-time geometry and Orch OR theory. J. Cosmol. 14, 1–17.

Google Scholar

Petroni, F., Rocktäschel, T., Lewis, P., Bakhtin, A., Wu, Y., Miller, A. H., et al. (2019). Language models as knowledge bases? arXiv [Preprint]. arXiv:1909.01066. (Accessed October 12, 2025).

Google Scholar

Pfeifer, R., and Bongard, J. (2006). How the body shapes the way we think: A new view of intelligence. Cambridge, MA: MIT press.

Google Scholar

Pfeiffer, M., and Pfeil, T. (2018). Deep learning with spiking neurons: opportunities and challenges. Front. Neurosci. 12:774. doi: 10.3389/fnins.2018.00774,

PubMed Abstract | Crossref Full Text | Google Scholar

Pockett, S. (2000). The nature of consciousness: A hypothesis : iUniverse.

Google Scholar

Posner, M. I., and Petersen, S. E. (1990). The attention system of the human brain. Annu. Rev. Neurosci. 13, 25–42. doi: 10.1146/annurev.ne.13.030190.000325,

PubMed Abstract | Crossref Full Text | Google Scholar

Potter, S. M., El Hady, A., and Fetz, E. E. (2014). Closed-loop neuroscience and neuroengineering. Front. Neural Circuits 8:115. doi: 10.3389/fncir.2014.00115,

PubMed Abstract | Crossref Full Text | Google Scholar

Preskill, J. (2018). Quantum computing in the NISQ era and beyond. Quantum 2:79. doi: 10.22331/q-2018-08-06-79

Crossref Full Text | Google Scholar

Qadri, R., Havaei, D., Denton, E., Al-Halah, Z., Diaz, F., Rostamzadeh, N., et al. (2022). Ethical considerations in creative applications of computer vision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 1321–1330). Available online at: https://cvpr.thecvf.com/virtual/2023/workshop/18488 (Accessed October 12, 2025).

Google Scholar

Qian, X., Song, H., and Ming, G. L. (2020). Brain organoids: advances, applications and challenges. Development 147:dev166074.

Google Scholar

Qiu, X., Sun, T., Xu, Y., Shao, Y., Dai, N., and Huang, X. (2020). Pre-trained models for natural language processing: a survey. Sci. China Technol. Sci. 63, 1872–1897. doi: 10.1007/s11431-020-1647-3

Crossref Full Text | Google Scholar

Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., et al. 2021. Learning transferable visual models from natural language supervision. In International conference on machine learning (pp. 8748–8763).

Google Scholar

Raji, I. D., Scheuerman, M. K., and Amironesei, R. 2021. You can't sit with us: exclusionary pedagogy in AI ethics education. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (pp. 515–525).

Google Scholar

Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., and Chen, M. (2022). Hierarchical text-conditional image generation with clip latents. arXiv [Preprint]. arXiv:2204.06125.

Google Scholar

Reardon, S. (2020). Can lab-grown brains become conscious? Nature 586, 658–661. doi: 10.1038/d41586-020-02986-y,

PubMed Abstract | Crossref Full Text | Google Scholar

Reed, S., Zolna, K., Parisotto, E., Colmenarejo, S. G., Novikov, A., Barth-Maron, G., et al. (2022). A generalist agent. Transact. Mach. Learn. Res. arXiv:2205.06175. doi: 10.48550/arXiv.2205.06175

Crossref Full Text | Google Scholar

Reggia, J. A. (2013). The rise of machine consciousness: studying consciousness with computational models. Neural Netw. 44, 112–131. doi: 10.1016/j.neunet.2013.03.011,

PubMed Abstract | Crossref Full Text | Google Scholar

Ritchie, G. (2007). Some empirical criteria for attributing creativity to a computer program. Minds Mach. 17, 67–99. doi: 10.1007/s11023-007-9066-2

Crossref Full Text | Google Scholar

Rogers, A., Kovaleva, O., and Rumshisky, A. (2020). A primer in BERTology: what we know about how BERT works. Trans. Assoc. Comput. Linguist. 8, 842–866 (If intending the neural network primer, change to: Goldberg, Y. (2017). A primer on neural network models for natural language processing. Journal of Artificial Intelligence Research, 58, 345-420.). doi: 10.1162/tacl_a_00349

Crossref Full Text | Google Scholar

Rosanova, M., Gosseries, O., Casarotto, S., Boly, M., Casali, A. G., Bruno, M. A., et al. (2012). Recovery of cortical effective connectivity and recovery of consciousness in vegetative patients. Brain 135, 1308–1320. doi: 10.1093/brain/awr340,

PubMed Abstract | Crossref Full Text | Google Scholar

Rosenthal, D. M. (2005). Consciousness and mind : Oxford University Press.

Google Scholar

Roy, K., Jaiswal, A., and Panda, P. (2019). Towards spike-based machine intelligence with neuromorphic computing. Nature 575, 607–617. doi: 10.1038/s41586-019-1677-2,

PubMed Abstract | Crossref Full Text | Google Scholar

Runco, M. A., and Jaeger, G. J. (2012). The standard definition of creativity. Creat. Res. J. 24, 92–96. doi: 10.1080/10400419.2012.650092

Crossref Full Text | Google Scholar

Russell, S. (2019). Human compatible: Artificial intelligence and the problem of control : Viking.

Google Scholar

Russell, S., and Norvig, P. (2020). Artificial intelligence: A modern approach : Pearson.

Google Scholar

Said-Metwaly, S., Noortgate, W. V. D., and Kyndt, E. (2017). Approaches to measuring creativity: a systematic literature review. Creativity Theor. Res. Appl. 4, 238–275. doi: 10.1515/ctra-2017-0013

Crossref Full Text | Google Scholar

Sandberg, A., and Bostrom, N. (2008). Whole brain emulation: A roadmap : Oxford University, 1–130.

Google Scholar

Sarasso, S., Boly, M., Napolitani, M., Gosseries, O., Charland-Verville, V., Casarotto, S., et al. (2015). Consciousness and complexity during unresponsiveness induced by propofol, xenon, and ketamine. Curr. Biol. 25, 3099–3105. doi: 10.1016/j.cub.2015.10.014,

PubMed Abstract | Crossref Full Text | Google Scholar

Saunders, W., Yeh, C., Wu, J., Bills, S., Ouyang, L., Ward, J., et al. (2022). Self-critiquing models for assisting human evaluators. arXiv [Preprint]. arXiv:2206.05802.

Google Scholar

Schacter, D. L., and Addis, D. R. (2007). The cognitive neuroscience of constructive memory: remembering the past and imagining the future. Philos. Trans. R. Soc. B 362, 773–786. doi: 10.1098/rstb.2007.2087,

PubMed Abstract | Crossref Full Text | Google Scholar

Schlosshauer, M. (2019). Quantum decoherence. Phys. Rep. 831, 1–57. doi: 10.1016/j.physrep.2019.10.001

Crossref Full Text | Google Scholar

Schneider, S. (2019). Artificial you: AI and the future of your mind. Princeton, NJ: Princeton University Press.

Google Scholar

Schreiber, T. (2000). Measuring information transfer. Phys. Rev. Lett. 85, 461–464. doi: 10.1103/PhysRevLett.85.461,

PubMed Abstract | Crossref Full Text | Google Scholar

Schuman, C. D., Potok, T. E., Patton, R. M., Birdwell, J. D., Dean, M. E., Rose, G. S., et al. (2017). A survey of neuromorphic computing and neural networks in hardware. arXiv [Preprint]. arXiv:1705.06963.

Google Scholar

Sergent, C., and Dehaene, S. (2004). Is consciousness a gradual phenomenon? Evidence for an all-or-none bifurcation during the attentional blink. Psychol. Sci. 15, 720–728. doi: 10.1111/j.0956-7976.2004.00748.x,

PubMed Abstract | Crossref Full Text | Google Scholar

Seth, A. K. (2016). The real problem of consciousness. Aeon Magazine. Available at: https://aeon.co/essays/the-hard-problem-of-consciousness-is-a-distraction-from-the-real-one

Google Scholar

Seth, A. K., and Bayne, T. (2022). Theories of consciousness. Nat. Rev. Neurosci. 23, 439–452. doi: 10.1038/s41583-022-00587-4,

PubMed Abstract | Crossref Full Text | Google Scholar

Seth, A. K., Dienes, Z., Cleeremans, A., Overgaard, M., and Pessoa, L. (2008). Measuring consciousness: relating behavioural and neurophysiological approaches. Trends Cogn. Sci. 12, 314–321. doi: 10.1016/j.tics.2008.04.008,

PubMed Abstract | Crossref Full Text | Google Scholar

Seth, A. K., Suzuki, K., and Critchley, H. D. (2012). An interoceptive predictive coding model of conscious presence. Front. Psychol. 2:395. doi: 10.3389/fpsyg.2011.00395,

PubMed Abstract | Crossref Full Text | Google Scholar

Shanahan, M. (2006). A cognitive architecture that combines internal simulation with a global workspace. Conscious. Cogn. 15, 433–449. doi: 10.1016/j.concog.2005.11.005,

PubMed Abstract | Crossref Full Text | Google Scholar

Shen, Y., Harris, N. C., Skirlo, S., Prabhu, M., Baehr-Jones, T., Hochberg, M., et al. (2017). Deep learning with coherent nanophotonic circuits. Nat. Photonics 11, 441–446. doi: 10.1038/nphoton.2017.93

Crossref Full Text | Google Scholar

Shenhav, A., Botvinick, M. M., and Cohen, J. D. (2013). The expected value of control: an integrative theory of anterior cingulate cortex function. Neuron 79, 217–240. doi: 10.1016/j.neuron.2013.07.007,

PubMed Abstract | Crossref Full Text | Google Scholar

Shoeybi, M., Patwary, M., Puri, R., LeGresley, P., Casper, J., and Catanzaro, B. (2019). Megatron-lm: training multi-billion parameter language models using model parallelism. arXiv [Preprint]. arXiv:1909.08053. (Accessed October 12, 2025).

Google Scholar

Shrestha, S. B., and Orchard, G. (2018). Slayer: Spike layer error reassignment in time. Adv. Neural Inf. Proces. Syst. 31.

Google Scholar

Sigman, M., and Dehaene, S. (2008). Brain mechanisms of serial and parallel processing during dual-task performance. J. Neurosci. 28, 7585–7598. doi: 10.1523/JNEUROSCI.0948-08.2008,

PubMed Abstract | Crossref Full Text | Google Scholar

Signorelli, C. M., Szczotka, J., and Prentner, R. (2021). Explanatory profiles of models of consciousness-towards a systematic classification. Neurosci. Conscious. 2021. doi: 10.1093/nc/niab021,

PubMed Abstract | Crossref Full Text | Google Scholar

Silver, D., Huang, A., Maddison, C. J., Guez, A., Sifre, L., Van Den Driessche, G., et al. (2016). Mastering the game of go with deep neural networks and tree search. Nature 529, 484–489. doi: 10.1038/nature16961,

PubMed Abstract | Crossref Full Text | Google Scholar

Simian, M., and Bissell, M. J. (2017). Organoids: a historical perspective of thinking in three dimensions. J. Cell Biol. 216, 31–40. doi: 10.1083/jcb.201610056,

PubMed Abstract | Crossref Full Text | Google Scholar

Simon, H. A. (1973). The structure of ill structured problems. Artif. Intell. 4, 181–201. doi: 10.1016/0004-3702(73)90011-8

Crossref Full Text | Google Scholar

Skinner, B. F. (1948). 'Superstition' in the pigeon. J. Exp. Psychol. 38, 168–172. doi: 10.1037/h0055873,

PubMed Abstract | Crossref Full Text | Google Scholar

Sloman, S. A., and Lagnado, D. (2015). Causality in thought. Annu. Rev. Psychol. 66, 223–247. doi: 10.1146/annurev-psych-010814-015135,

PubMed Abstract | Crossref Full Text | Google Scholar

Smirnova, L., Caffo, B. S., Gracias, D. H., Huang, Q., Igel, C., Kozloski, J., et al. (2023). Organoid intelligence (OI): the new frontier in biocomputing and intelligence-in-a-dish. Front. Sci. 1:1017235.

Google Scholar

Soares, N., and Fallenstein, B. (2017). “Agent foundations for aligning machine intelligence with human interests: a technical research agenda” in The technological singularity (Princeton NJ: Springer), 103–125.

Google Scholar

Squire, L. R. (2004). Memory systems of the brain: a brief history and current perspective. Neurobiol. Learn. Mem. 82, 171–177. doi: 10.1016/j.nlm.2004.06.005,

PubMed Abstract | Crossref Full Text | Google Scholar

Squire, L. R., and Kandel, E. R. (2009). Memory: From mind to molecules : Scientific American Library.

Google Scholar

Strubell, E., Ganesh, A., and McCallum, A. 2019. Energy and policy considerations for deep learning in NLP. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 3645–3650.

Google Scholar

Suzgun, M., Scales, N., Schärli, N., Gehrmann, S., Tay, Y., Chung, H. W., et al. (2022). Challenging BIG-bench tasks and whether chain-of-thought can solve them. arXiv [Preprint]. arXiv:2210.09261. (Accessed October 12, 2025).

Google Scholar

Sze, V., Chen, Y. H., Yang, T. J., and Emer, J. S. (2017). Efficient processing of deep neural networks: a tutorial and survey. Proc. IEEE 105, 2295–2329. doi: 10.1109/JPROC.2017.2761740

Crossref Full Text | Google Scholar

Tagliazucchi, E., and Laufs, H. (2014). Decoding wakefulness levels from typical fMRI resting-state data reveals reliable drifts between wakefulness and sleep. Neuron 82, 695–708. doi: 10.1016/j.neuron.2014.03.020,

PubMed Abstract | Crossref Full Text | Google Scholar

Tegmark, M. (2000). Importance of quantum decoherence in brain processes. Phys. Rev. E 61, 4194–4206. doi: 10.1103/PhysRevE.61.4194,

PubMed Abstract | Crossref Full Text | Google Scholar

Tenney, I., Das, D., and Pavlick, E. (2019). BERT rediscovers the classical NLP pipeline. arXiv [Preprint]. arXiv:1905.05950. (Accessed October 12, 2025).

Google Scholar

Thagard, P. (1988). Computational philosophy of science. Cambridge, MA: MIT press.

Google Scholar

Tononi, G. (2008). An information integration theory of consciousness. BMC Neurosci. 9, 1–22.

Google Scholar

Tononi, G., Boly, M., Massimini, M., and Koch, C. (2016). Integrated information theory: from consciousness to its physical substrate. Nat. Rev. Neurosci. 17, 450–461. doi: 10.1038/nrn.2016.44,

PubMed Abstract | Crossref Full Text | Google Scholar

Tulving, E. (1972). “Episodic and semantic memory” in Organization of memory (San Diego: Academic Press), 381–403.

Google Scholar

UNESCO (2021) in Recommendation on the ethics of artificial intelligence. ed. UNESCO (Paris).

Google Scholar

VanRullen, R., and Koch, C. (2003). Is perception discrete or continuous? Trends Cogn. Sci. 7, 207–213. doi: 10.1016/S1364-6613(03)00095-0,

PubMed Abstract | Crossref Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). Attention is all you need. Adv. Neural Inf. Proces. Syst. 30.

Google Scholar

Veale, T., and O'Donoghue, D. (2000). Computation and blending. Cogn. Linguist. 11, 253–281.

Google Scholar

Wang, J. X., Kurth-Nelson, Z., Tirumala, D., Soyer, H., Leibo, J. Z., Munos, R., et al. (2022). Learning to reinforcement learn. arXiv [Preprint]. arXiv:1611.05763. (Accessed October 12, 2025).

Google Scholar

Ward, L. M. (2011). The thalamic dynamic core theory of conscious experience. Conscious. Cogn. 20, 464–486. doi: 10.1016/j.concog.2011.01.007,

PubMed Abstract | Crossref Full Text | Google Scholar

Wei, J., Tay, Y., Bommasani, R., Raffel, C., Zoph, B., Borgeaud, S., et al. (2022). Emergent abilities of large language models. Transact. Mach. Learn. Res. doi: 10.48550/arXiv.2206.0768

Crossref Full Text | Google Scholar

Wetzstein, G., Ozcan, A., Gigan, S., Fan, S., Englund, D., Soljačić, M., et al. (2020). Inference in artificial intelligence with deep optics and photonics. Nature 588, 39–47. doi: 10.1038/s41586-020-2973-6,

PubMed Abstract | Crossref Full Text | Google Scholar

Wiese, W., and Friston, K. J. (2021). Examining the continuity between life and mind: is there a continuity between autopoietic intentionality and representationality? Philosophies 6:18. doi: 10.3390/philosophies6010018

Crossref Full Text | Google Scholar

Wiggins, G. A. (2006). A preliminary framework for description, analysis and comparison of creative systems. Knowl.-Based Syst. 19, 449–458. doi: 10.1016/j.knosys.2006.04.009

Crossref Full Text | Google Scholar

Wiggins, G. A., and Bhattacharya, J. (2014). Mind the gap: an attempt towards a model of creative insight. Front. Hum. Neurosci. 8:540.

Google Scholar

Williford, K., Bennequin, D., Friston, K., and Rudrauf, D. (2018). The projective consciousness model and phenomenal selfhood. Front. Psychol. 9:2571. doi: 10.3389/fpsyg.2018.02571,

PubMed Abstract | Crossref Full Text | Google Scholar

Wilson, M. (2002). Six views of embodied cognition. Psychon. Bull. Rev. 9, 625–636. doi: 10.3758/BF03196322,

PubMed Abstract | Crossref Full Text | Google Scholar

Wittek, P. (2014). Quantum machine learning: What quantum computing means to data mining : Academic Press.

Google Scholar

Yoshida, M., Hirose, K., Yamamoto, N., and Moriya, H. (2021). Context-dependent modulation of consciousness through the global neuronal workspace. Neurosci. Res. 171, 12–23.

Google Scholar

Yu, J., Wang, Z., Vasudevan, V., Yeung, L., Seyedhosseini, M., and Wu, Y. (2022). CoCa: Contrastive captioners are image-text foundation models. arXiv [Preprint]. arXiv:2205.01917. (Accessed October 12, 2025).

Google Scholar

Zarkov, S., Kleiner, J., and Serre, T. (2024). Assessing machine consciousness via cross-domain metric integration. Conscious. Cogn. 118:103642.

Google Scholar

Zhou, Z., and Montague, P. R. (2017). Computational psychiatry: a new perspective on mental illness. Cerebrum 2017:cer-04-17.

Google Scholar

Keywords: machine consciousness, artificial intelligence, creativity, neuromorphic computing, organoid intelligence, perturbational complexity, agency, evaluation frameworks

Citation: Niazi SK (2026) Beyond mimicry: a framework for evaluating genuine intelligence in artificial systems. Front. Artif. Intell. 8:1686752. doi: 10.3389/frai.2025.1686752

Received: 15 August 2025; Revised: 12 November 2025; Accepted: 09 December 2025;
Published: 12 January 2026.

Edited by:

Ramesh Chandra Poonia, Christ University, India

Reviewed by:

Mohammad Ali Yamin, Jeddah University, Saudi Arabia
Nicolas Rouleau, Wilfrid Laurier University, Canada

Copyright © 2026 Niazi. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Sarfaraz K. Niazi, c25pYXppM0B1aWMuZWR1

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.