Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Artif. Intell., 20 January 2026

Sec. Machine Learning and Artificial Intelligence

Volume 9 - 2026 | https://doi.org/10.3389/frai.2026.1714523

This article is part of the Research TopicAdvancing AI-Driven Code Generation and Synthesis: Challenges, Metrics, and Ethical ImplicationsView all 4 articles

AI-assisted design synthesis and human creativity in engineering education

Mariza TsakalerouMariza TsakalerouSaltanat Akhmadi
Saltanat Akhmadi*Aruzhan BalgynbayevaAruzhan BalgynbayevaYerdaulet KumisbekYerdaulet Kumisbek
  • Department of Civil and Environmental Engineering, Nazarbayev University, Astana, Kazakhstan

The growing integration of AI into educational and professional settings raises urgent questions about how human creativity evolves when intelligent systems guide, constrain, or accelerate the design process. Generative AI offers structured suggestions and rapid access to ideas, but its role in adopting genuine innovation remains contested. This paper investigates the dynamics of human-AI collaboration in challenge-based design experiments, applying established creativity metrics: fluency, flexibility, originality, and elaboration in order to evaluate outcomes and implications in an engineering education context. Through an exploratory quasi-experimental study, a comparison of AI-assisted and human-only teams was conducted across four dimensions of creative performance: quantity, variety, uniqueness, and quality of design solutions. Findings point to a layered outcome: although AI accelerated idea generation, it also encouraged premature convergence, narrowed exploration, and compromised functional refinement. Human-only teams engaged in more iterative experimentation and produced designs of higher functional quality and greater ideational diversity. Participants’ self-perceptions of creativity remained stable across both conditions, highlighting the risk of cognitive offloading, where reliance on AI may reduce genuine creative engagement while masking deficits through inflated confidence. Importantly, cognitive offloading is not directly measured in this study; rather, it is introduced here as a theoretically grounded interpretive explanation that helps contextualize the observed disconnect between performance outcomes and self-perceived creativity. These results bring opportunities and risks. On the one hand, AI can support ideation and broaden access to concepts; on the other, overreliance risks weakening iterative learning and the development of durable creative capacities. The ethical implications are significant, raising questions about accountability and educational integrity when outcomes emerge from human-AI co-creation. The study argues for process-aware and ethically grounded frameworks that balance augmentation with human agency, supporting exploration without eroding the foundations of creative problem-solving. The study consolidates empirical findings with conceptual analysis, advancing the discussion on when and how AI should guide the creative process and providing insights for the broader debate on the future of human–AI collaboration.

1 Introduction

1.1 Generation, synthesis and creativity

The emergence of AI-driven code generation and synthesis technologies has revolutionized how creative problem-solving is approached in software development and engineering design. These systems demonstrate unprecedented capabilities in combining existing knowledge patterns to produce novel code solutions, design concepts, and technical implementations. However, as AI tools become increasingly sophisticated at both creating new content and incorporating disparate information sources, fundamental questions arise about their impact on human creative processes.

Modern AI systems excel at synthesis, which the Cambridge Dictionary defines as “the mixing of different ideas, influences, or things to make a whole that is different, or new.” This capability allows them to systematically recombine existing knowledge from vast repositories, mirroring how advanced code-generation tools integrate patterns, functions, and algorithmic structures to create novel implementations (Haase and Hanel, 2023; Kirkpatrick, 2023). However, the distinction between this sophisticated recombination and genuine creativity, defined as “the ability to produce or use original and unusual ideas,” remains contested.

While AI systems demonstrate remarkable capabilities in combining knowledge in increasingly sophisticated ways, their capacity for true innovation is fundamentally constrained by their reliance on pre-existing data (Habib et al., 2024). This limitation becomes particularly significant in engineering education, where students must master both the effective use of existing solutions and the development of genuinely innovative approaches to novel problems. Some scholars argue that even AI’s ability to integrate remote and unrelated knowledge domains enables innovative outcomes that produce “original and unused ideas” (Lee and Chung, 2024), suggesting that cross-domain integration capabilities have important implications for both code generation and broader engineering design processes.

Foundational work by Guilford (1950) and Torrance (1966) emphasizes that creative performance depends on divergent thinking processes, specifically the fluency, flexibility, and originality of ideas generated. These classic insights examine whether AI-assisted teams engage in the expansive, exploratory ideation required for genuine creative development. Understanding how these AI capabilities affect human creative development requires careful examination, particularly in educational contexts where the goal extends beyond producing outputs to developing students’ own creative problem-solving abilities.

1.2 Human and AI creativity

The tension between AI capabilities and human creativity has led to extensive empirical research comparing these abilities using established measures such as the Alternative Uses Task (AUT) and Torrance tests (Góes et al., 2023; Stevenson et al., 2022; Summers-Stay et al., 2023). Moreover, empirical findings have prompted researchers to develop new terminology describing different forms of AI-involved creativity, reflecting the evolving relationship between human designers and intelligent systems: artificial creativity (Ivcevic and Grandinetti, 2024; Runco, 2023), computational creativity (or creative computation) (Colton and Wiggins, 2012), and assisted creativity or co-creativity (Wingström et al., 2024).

The concept of assisted creation processes is particularly relevant to code generation and design synthesis. Recent research suggests we are entering “a new era of ‘assisted creativity,’ where AI serves not as an independent creator but as a collaborative creative agent” (Habib et al., 2024; Vinchon et al., 2023). This collaborative model reflects how modern programming increasingly involves human-AI partnerships, where developers guide intelligent tools to generate code meeting specific requirements while adding their own insights to solve complex problems. At the same time, prior work in engineering education has shown that while AI tools are perceived by both students and faculty as valuable for explanation, content generation, and learning support, they also raise concerns related to academic integrity, overreliance, and diminished critical engagement (Lukhmanov et al., 2025).

Recent philosophical work further clarifies the limits of AI-assisted creativity by foregrounding the ontological boundaries of machine generation. Lockhart (2025) argues that while AI systems can simulate novelty through recombination, creativity remains fundamentally human because it is grounded in embodied experience, emotional depth, and moral agency. From this perspective, AI-generated outputs lack the intentionality, self-authorship, and existential risk-taking that characterize human creative acts. Lockhart’s framework complements empirical models of assisted and co-creativity (Ivcevic and Grandinetti, 2024; Wingström et al., 2024) by emphasizing that augmentation does not imply equivalence. In educational contexts, this distinction is particularly important: when AI tools guide ideation without reflective engagement, learners may produce technically novel outputs while disengaging from the cognitive and ethical processes that sustain authentic creative development. This human-centered perspective provides a conceptual lens for interpreting empirical patterns such as premature convergence and cognitive offloading observed in AI-assisted design teams, reinforcing the need for pedagogical frameworks that preserve human agency alongside technological support.

These collaborative frameworks have significant implications for engineering education, where students must learn to work effectively with intelligent systems while developing their own problem-solving capabilities. Following Haase and Hanel (2023), while philosophical debates about the nature of innovation continue, the more pressing educational question concerns how widespread AI use affects human creative development in practical design and programming contexts.

1.3 Creativity metrics and measurement approaches

Given the evolving nature of human-AI collaboration in creative tasks, establishing reliable assessment metrics becomes crucial for understanding educational impacts. Design and engineering contexts have developed sophisticated approaches applicable to both traditional and AI-assisted creative processes.

The most widely accepted framework evaluates innovation across two primary dimensions: novelty and usefulness (Amabile, 1982; Doshi and Hauser, 2024; Harvey and Berry, 2022). This dual-dimension approach proves particularly relevant for AI-assisted contexts, as it separates idea generation from practical application, paralleling the difference between AI’s generative capabilities and human evaluative judgment. Novelty assesses how much an idea departs from expectations, while usefulness reflects practicality and relevance, including appropriateness, feasibility, and implementability (Doshi and Hauser, 2024). These dimensions translate effectively into code-generation contexts, where novel algorithmic approaches must meet functional requirements and coding standards.

Building upon this foundation, researchers have also adapted classical creativity metrics originally developed by Torrance (1966) for assessing human creative thinking:

• fluency (quantity of ideas),

• flexibility (diversity or variety of categories of ideas),

• originality (uncommonness or uniqueness of ideas), and

• elaboration (development depth or quality of ideas).

As shown in Table 1, among the various dimensions employed in recent empirical studies, four metrics appear most frequently across different task types and provide comprehensive coverage of creative output assessment.

Table 1
www.frontiersin.org

Table 1. Overview of recent empirical studies on the human and AI creativity.

Notably, recent empirical studies do not limit the application of these metrics to human creativity assessment alone but expand their scope to evaluate AI-generated outputs, human-AI collaborative processes, and comparative analyses between traditional and AI-assisted creative work (Table 1). Studies examining human and AI creativity demonstrate considerable variation in which dimensions are measured and how they are operationalized across different task types and populations (Guzik et al., 2023; Habib et al., 2024; Marrone et al., 2024). This heterogeneity in measurement approaches reflects both the evolving nature of the field and ongoing debates about appropriate assessment frameworks for AI-augmented creative work.

However, AI-augmented contexts introduce new complexities. As Wingström et al. (2024) note, creative AI “contests the issues regarding novelty, autonomy, and authorship, as AI’s creativity is often evaluated via the outcome it produces.” These concerns become particularly acute in educational settings, where goals extend beyond producing outputs to developing students’ own creative capabilities.

1.4 AI integration in engineering education: opportunities and challenges

The integration of AI tools in engineering education reflects broader global trends, with countries such as Singapore, Estonia, Australia, New Zealand, and Scotland leading implementation efforts (Gabriel et al., 2022; Marrone et al., 2022). These initiatives recognize that preparing students for AI-augmented professional environments requires hands-on experience during education. However, this transformation raises critical questions about pedagogical approaches, particularly regarding prompt engineering skills and problem-definition capabilities, which remain essential in both code-generation and design contexts (Haase and Hanel, 2023).

Despite extensive research on AI creativity and growing educational interest, a significant gap remains in understanding how AI usage impacts human creativity in collaborative educational settings. Most existing studies focus on philosophical questions about whether AI can be creative, but fewer examine the practical question of how AI assistance affects human creative development in team-based contexts. This gap is particularly pronounced in engineering education research, where design synthesis and creative problem-solving represent fundamental skills students must master.

As Haase and Hanel (2023) note, “Continued research and development of GAI [Generative AI] in creative tasks is crucial to fully understand this technology’s potential benefits and drawbacks in shaping the future of creativity.” The limited research on team-based AI-assisted creativity leaves important questions unanswered: How do AI tools affect collaborative creative processes that characterize real-world engineering practice? Do AI-assisted teams produce more creative solutions, or does AI assistance reduce individual creative development? Moreover, the parallel between design creativity and programming creativity suggests that findings from design-focused research may inform broader questions about AI’s impact on creative problem-solving across technical domains.

To address these critical gaps, this exploratory study investigates two interconnected research questions:

RQ1 examines how AI assistance affects team creativity across four established dimensions and identifies which aspects of creativity are most influenced by AI usage. Specifically:

• RQ1a: Do AI-assisted teams differ from control teams in creative output across the four dimensions?

• RQ1b: Which dimension of creativity shows the most significant difference between AI-assisted and non-AI-assisted conditions?

RQ2 examines educational implications by analyzing participants’ self-perceptions of creative capability, exploring how AI assistance affects individual confidence in creative problem-solving and its implications for educational practice. Specifically:

• RQ2a: Does AI assistance lead students to overestimate their creative performance compared to objective creativity metrics?

• RQ2b: How does this perception–performance discrepancy differ between AI-assisted and non-AI teams?

This study addresses the research gaps through a systematic comparative investigation contrasting AI-assisted teams with human-driven teams in collaborative engineering design tasks. Our approach builds on established creativity assessment frameworks, specifically adapted for educational contexts, to evaluate team performance across four dimensions that capture both divergent thinking aspects (Quantity, Variety) and convergent thinking aspects (Quality), with Uniqueness bridging generative and evaluative processes. Through this comparative evaluation, the research explores the cognitive, functional, and ethical implications of AI-assisted design, contributing to understanding not only how AI affects team creativity but also broader questions about AI-driven synthesis and automated problem-solving—the issues central to the evolving relationship between human creativity and artificial intelligence in technical domains.

2 Materials and methods

2.1 Study design and experimental conditions

This exploratory, comparative, quasi-experimental investigation was embedded within a 90-min workshop focused on creative engineering practices (Photographic documentation of the workshop set-up and environment is presented in Appendix A). The study employed a between-subjects design to examine how AI assistance influences team creativity in collaborative engineering design tasks, utilizing the Catapult Challenge adapted from Seelig’s “creativity-under-constraint” framework (Seelig, 2017).

A convenience cohort of volunteers, comprising postgraduate students, early-career lecturers, and industry mentors, self-organized into eight teams of two to three members each. Demographically, participants were predominantly engineering-focused (85% reported an Engineering background), 62% male and 38% female, with 30% under 34 years and 70% aged 45 or older, representing both emerging and established professionals. To ensure unbiased allocation, the session chair prepared 12 shuffled, opaque envelopes (seven designated as “AI group” and five designated as “non-AI group”), and immediately before the task, one representative from each self-organized team drew an envelope to determine random condition assignment. This allocation method was blind to participants’ background, experience, and prior exposure to design tasks, providing baseline equivalence across these variables. All participants had no prior exposure to the experimental task and were clearly informed of the AI usage restrictions applicable to their assigned condition.

The experimental task required teams to design and construct a functional tabletop catapult capable of launching a marble at least 50 centimeters within a 45-min timeframe. Each team received an identical construction kit containing wooden craft sticks, rubber bands, plastic spoons, masking tape, a marble, scissors, and a metric measuring tape. Teams assigned to the AI-assisted condition received uninterrupted browser access to ChatGPT-4 (web version 2025-03-13) along with a standardized prompt sheet that read: “Suggest three distinct catapult concepts using only craft sticks, rubber bands, and a spoon. For each concept, list the mechanical principle, a simple ASCII sketch, and one improvement tip.” These teams retained complete autonomy to refine, modify, or disregard ChatGPT’s suggestions according to their judgment. Control teams were permitted to sketch freely and consult non-LLM web resources such as Wikipedia or YouTube, but received explicit reminders that large language models were prohibited. No financial incentives were provided to ensure participation remained voluntary and intrinsically motivated.

2.2 Measurement framework and experimental procedure

The study employed both individual-level and team-level assessments to capture different dimensions of creative performance within a structured experimental protocol. The experimental session began with pre-survey completion (approximately 10 min), during which individual participants provided baseline measures of creative self-efficacy using a four-item, three- to five-point Likert scale, alongside single-item measures of divergent-thinking confidence, problem-solving efficacy, and innovative-thinking self-ratings. These instruments were adapted from the 4C Creativity Assessment and administered via Google Forms. In the scope of this study, Little-c or everyday creativity was addressed, aiming to assess “original and appropriate ideas or products in the context of everyday life and interactions” (Ivcevic and Grandinetti, 2024).

Following a five-minute briefing that included specific AI-usage guidelines for applicable teams, participants engaged in the 45-min Catapult Challenge with unrestricted access to a designated testing station for prototype evaluation. Upon task completion, all participants immediately completed the post-survey (approximately 10 min), which included team-level creativity assessments using the four-dimensional framework synthesized from the metrics established in the literature review:

1. Quantity (fluency)—“How many catapult designs did your team consider?” (5-point Likert scale ranging from 1 design to 5+ designs).

2. Variety (flexibility)—“Did designs employ different mechanical principles?” (3-point Likert scale: 0 = no variety, 1 = medium variety, 2 = distinct variety).

3. Uniqueness (originality/novelty)—“How unique was your final design?” (3-point Likert scale: 0 = standard design, 1 = somewhat unique design, 2 = unique design).

4. Quality (usefulness)—“How well did your catapult work?” (3-point Likert scale: 0 = failed, 1 = worked somewhat, 2 = worked well).

Two independent raters transferred categorical responses to a master coding sheet, achieving strong inter-coder agreement (κ = 0.94). Both surveys are presented in Appendix B.

The session concluded with a brief, ungraded showcase and debrief discussion, though no feedback from this final stage was incorporated into the dataset to maintain the integrity of the collected measures.

2.3 Data analysis

All analyses were conducted using R 4.3 software. Data were exported to CSV format, screened for missing values (less than 2% across all items), and ordinal-coded for statistical analysis. Descriptive statistics, including means with standard deviations and frequency percentages, summarized individual and team outcomes. Given the small cell counts inherent to the experimental design, Mann–Whitney U tests were used to assess differences in post-survey creativity scores between AI-assisted and control teams. Pre-to-post changes within each cohort were evaluated using Wilcoxon signed-rank tests. Statistical significance was established at α = 0.05 using two-tailed tests, with no correction for multiple comparisons given the exploratory nature of the investigation.

3 Results

All analyses were conducted at two complementary levels: team-level design outcomes (Quantity, Variety, Uniqueness, and Quality) and individual-level creative self-ratings recorded before and after the Catapult Challenge. The sample size is AI = 7 teams; non-AI = 5 teams.

3.1 Team-level design outcomes

Mann–Whitney U tests revealed no statistically significant differences between AI-assisted and non-AI teams at α = 0.05. However, effect sizes ranged from small to large (r = 0.26–0.56), with Uniqueness approaching significance and the remaining three showing a trend. The consistent direction of effects suggests non-AI teams outperformed AI-assisted teams across all measures, though statistical power was limited by the small sample size. (Mann–Whitney U tests did not reveal statistically significant differences between conditions; this outcome should be interpreted in light of the very small number of teams).

A further descriptive-comparative analysis also revealed consistent patterns favoring non-AI teams across multiple creativity dimensions. As illustrated in Figure 1, design exploration, or the Quantity, differed markedly between conditions. Non-AI teams demonstrated greater design breadth, with 60% (3 out of 5) considering three or more design variants compared to 0% of AI-assisted teams (U = 6 with p = 0.073 and effect size r = −0.52). This pattern suggests, but does not clearly demonstrate, that AI assistance may have encouraged premature convergence on initial concepts rather than supporting divergent exploration.

Figure 1
Bar chart comparing the percentage of designs created by AI teams and non-AI teams. AI teams created 0.0% for one design, 85.7% for two designs, and 0.0% for others. Non-AI teams created 14.3% for one design, 40.0% for two and three designs, 20.0% for four designs, and 0.0% for five or more designs.

Figure 1. Number of catapult designs considered by AI-assisted and non-AI control teams.

Figure 2 presents the Variety and Uniqueness outcomes, revealing differences in creative processes. While mechanical variety of designs showed modest differences between conditions (40% vs. 29% achieving distinct principles with U = 11.5, p = 0.373, and r = −0.26), perceived design originality favored non-AI teams, with 40% rating their final designs as “very unique” compared to 14% (1 out of 7) of AI teams (U = 8.5, p = 0.168, and r = −0.40). These directional patterns align with the broader trend favoring human-driven creativity.

Figure 2
Two bar charts comparing design variety and uniqueness for AI and non-AI teams. The left chart shows non-AI teams with 40% no variety, 20% medium variety, and 40% distinct variety. AI teams have 57.1% no variety and 42.9% distinct variety. The right chart shows non-AI teams with 20% standard, 40% somewhat unique, and AI teams with 71.4% standard, 14.3% somewhat unique, and 14.3% unique designs.

Figure 2. Variety of the designs considered, and uniqueness of the final design selected by AI-assisted and non-AI control teams.

The most pronounced difference emerged in functional performance, as depicted in Figure 3. All non-AI teams (100%) achieved catapults that “worked well,” while only 29% (2 out of 7) of AI-assisted teams reached this performance threshold (U = 5.0, p = 0.051, and r = −0.56). This substantial gap suggests that reliance on AI assistance may have compromised the iterative refinement process essential for achieving functional design solutions.

Figure 3
Bar chart comparing the quality of final designs between Non-AI teams and AI teams. Both team types had zero percent under

Figure 3. Quality of the final design assessed by AI-assisted and non-AI control teams.

Figure 4 provides an integrated perspective through radar visualization, illustrating how non-AI teams consistently outperformed AI-assisted teams across the four creativity dimensions. The pattern reveals a particularly stark contrast in functional quality and design exploration, with more modest differences in variety and uniqueness measures. The observed patterns are consistent with premature convergence, in the sense that AI-assisted teams explored fewer alternatives and converged earlier on initial concepts.

Figure 4
Radar chart comparing AI and Non-AI teams across four criteria: 3+ designs considered, distinct variety of designs, unique design selected, and design selected worked well. Non-AI teams (purple) outperform AI teams (pink) in all areas.

Figure 4. The four-dimensional framework for the team-level creativity assessments.

3.2 Individual-level self-perceptions

Individual participants’ pre- to post-task changes in creative self-efficacy, divergent-thinking confidence, problem-solving efficacy, and innovative-thinking ratings were minimal in both conditions (Wilcoxon |z| < 1.2, p > 0.20 for all comparisons). This lack of change in self-perceived creative capability suggests several critical educational implications. First, the brief AI-assisted intervention did not enhance participants’ confidence in their creative abilities, despite the availability of AI support tools. Conversely, participants in AI-assisted teams did not experience diminished self-efficacy despite their objectively lower performance on team creativity measures. This disconnect between actual performance (where non-AI teams significantly outperformed AI teams) and self-perceived capability raises questions about participants’ awareness of AI’s impact on their creative processes.

The stability of individual creative self-perceptions across conditions suggests that participants may not have fully recognized how AI assistance affected their collaborative creative processes. This finding has implications for educational practice, suggesting that students may require explicit instruction on the potential benefits and limitations of AI tools in creative contexts, rather than assuming they will naturally develop appropriate usage strategies through experience alone.

3.3 Qualitative observations

Workshop observations revealed behavioral differences that explain the quantitative results. AI-assisted teams spent substantial time formulating prompts, interpreting ChatGPT responses, and adapting suggested concepts to fit available materials. This process consumed time that could otherwise be used for physical testing and design refinement. Non-AI teams proceeded directly to building prototypes, allowing for more iterative testing and improvement cycles.

The time-allocation patterns suggest that AI interaction created additional cognitive demands that competed with hands-on problem-solving. While AI-assisted teams received conceptual guidance, they struggled to translate abstract suggestions into functional designs within the time constraints. Non-AI teams, working without external input, focused their efforts on direct experimentation with materials and mechanisms, resulting in more effective final products.

4 Future studies and limitations

This study employed a quasi-experimental design in which participants self-organized into teams before being randomly assigned to AI-assisted and non-AI conditions. The condition assignment was randomized at the team level, and baseline equivalence was examined. We acknowledge the absence of full individual-level randomization introduces the possibility of self-selection effects and unobserved confounding variables that may affect internal validity. Accordingly, the findings should not be interpreted as establishing a causal effect of AI assistance on creativity outcomes. Rather, they indicate systematic associations and comparative patterns between AI-assisted and human-only teams within the specific constraints of a short, challenge-based workshop setting. Future research using fully randomized designs, crossover experiments, or longitudinal approaches would be required to support stronger causal inferences.

A further methodological limitation concerns the reliance on self-reported measures of creativity. These instruments, which capture participants’ perceived creative capability and self-perception data, are susceptible to bias, including overconfidence or miscalibration, particularly in AI-assisted contexts where external support may mask performance deficits. In this context, future studies should triangulate self-assessments with expert or peer evaluations. In addition, several key explanatory constructs discussed in this study, such as premature convergence and cognitive offloading, are introduced as theoretically grounded interpretive mechanisms. It is important to operationalize these inferred mechanisms more directly by incorporating process-level methods, such as think-aloud protocols, systematic process tracing of design decisions, or fine-grained interaction logs capturing human–AI collaboration dynamics over time. Such approaches would allow closer examination of how AI tools shape ideation trajectories, iteration cycles, and decision-making processes during creative problem-solving.

Finally, the study captures human–AI collaboration at a specific point in the (very) rapid evolution of generative AI technologies and within a constrained time frame. Replication across different task durations, disciplinary contexts, and with newer generations of AI systems would help assess the robustness and generalizability of the observed patterns.

5 Discussion and conclusions

Before interpreting these findings, it is important to clarify the scope of inference permitted by the study design. Although teams were randomly assigned to AI-assisted and non-AI conditions, team formation occurred through self-organization prior to condition assignment. As a result, the quasi-experimental design does not support strong causal claims regarding the effects of AI assistance on creativity outcomes. The patterns discussed below should therefore be interpreted as comparative and associational, reflecting differences observed between AI-assisted and human-only teams under the specific constraints of a short, challenge-based workshop, rather than as evidence of direct causal effects.

Our findings suggest that AI assistance in team-based design tasks is associated with patterns consistent with premature convergence, leading to fewer explored alternatives and lower functional quality compared to human-only teams. While AI tools offered structured suggestions, they often constrained divergent exploration, limiting opportunities for iteration and refinement. This pattern is consistent with prior findings in AI-enhanced software development education, where large language models were found to accelerate early-stage ideation and task execution, while raising concerns about customization, transparency, and depth of engagement in later stages of development (Israilidis et al., 2024).

Cutting-edge research argues that current AI systems operate as fragmented utilities rather than process-aware collaborators (Wang and Lu, 2025). Their proposed layered framework emphasizes that effective human–AI collaboration requires explicit process representation and adaptability over time. Our results illustrate a pattern in which, when such support is absent, what happens when such support is absent: students rely on AI for isolated outputs but lose the iterative, hands-on engagement that supports deeper problem-solving.

At the same time, the Memory Paradox (Oakley et al., 2025) highlights how excessive reliance on external aids can erode internal cognitive development, weakening schema formation and procedural fluency. In our study, participants in AI-assisted teams maintained confidence in their creative abilities despite producing weaker outcomes, reflecting the risk that AI offloading may mask shallow engagement. Without building strong internal “creative schemata” through practice and reflection, learners will ultimately fail to develop transferable problem-solving skills. This perception–performance gap also highlights a methodological limitation or relying solely on self-reported measures, which may now capture actual creative performance.

The findings can be interpreted through a broader human-centered theoretical lens on creativity. Lockhart (2025) argues that creativity is not reducible to the generation of novel outputs but is fundamentally grounded in lived experience, embodied understanding, and moral agency. From this perspective, creative acts emerge through situated engagement with materials, constraints, and failure, rather than through abstract recombination alone.

Nevertheless, taken together, these observations point to a central tension: AI may assist in accelerating access to ideas, but without structured collaboration and deliberate memory-building, it risks narrowing exploration and undermining genuine learning. The challenge is not only to integrate AI responsibly but also to design environments that preserve human agency, iterative refinement, and the development of durable creative capacities.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors without undue reservation.

Ethics statement

The studies involving humans were approved by Ethical approval for this study was obtained on 18 October 2024. The study was reviewed in accordance with national and international ethical standards for research involving human participants. All participants were informed about the study’s purpose and procedures and provided written consent before participating. Data confidentiality, anonymity, and security were maintained throughout the project in accordance with GDPR and the national data protection laws of the Republic of Kazakhstan. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author contributions

MT: Writing – original draft, Writing – review & editing, Supervision, Funding acquisition, Formal analysis, Project administration, Conceptualization. SA: Writing – original draft, Methodology, Formal analysis, Data curation, Investigation. AB: Methodology, Writing – original draft, Data curation, Visualization, Investigation. YK: Formal analysis, Writing – original draft, Data curation, Visualization, Methodology.

Funding

The author(s) declared that financial support was received for this work and/or its publication. The research leading to these results received funding from Nazarbayev University under Grant Agreement. No. 201223FD8810 (FDCRG).

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that Generative AI was used in the creation of this manuscript. The authors declare that no generative AI tools were used to create the content, ideas, or theories presented in this manuscript. Under close human supervision, AI-assisted technologies were employed solely to enhance language clarity and readability. All AI-refined content was meticulously reviewed, edited, and validated by the authors to ensure accuracy and coherence. The authors acknowledge that while AI may assist in improving expression, it can also introduce errors or biases. Consequently, the final manuscript reflects human critical judgment and maintains scholarly integrity.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frai.2026.1714523/full#supplementary-material

References

Amabile, T. (1982). Social psychology of creativity: a consensual assessment technique. J. Pers. Soc. Psychol. 43, 997–1013. doi: 10.1037/0022-3514.43.5.997

Crossref Full Text | Google Scholar

Colton, S., and Wiggins, G. (2012). Computational creativity: the final frontier? Front. Art. Int. Appl. 242, 21–26. doi: 10.3233/978-1-61499-098-7-21

Crossref Full Text | Google Scholar

Doshi, A. R., and Hauser, O. P. (2024). Generative AI enhances individual creativity but reduces the collective diversity of novel content. Sci. Adv. 10, 1–9. doi: 10.1126/sciadv.adn5290,

PubMed Abstract | Crossref Full Text | Google Scholar

Gabriel, F., Marrone, R., Van Sebille, Y., Kovanovic, V., and Laat, M. (2022). Digital education strategies around the world: practices and policies. Ir. Educ. Stud. 41, 85–106. doi: 10.1080/03323315.2021.2022513

Crossref Full Text | Google Scholar

Góes, L. F., Volpe, M., Sawicki, P., Grses, M., and Watson, J. (2023). Pushing GPT’s creativity to its limits: alternative uses and torrance tests. Leicester: University of Leicester.

Google Scholar

Guilford, J. P. (1950). Creativity. Am. Psychol. 5, 444–454. doi: 10.1037/h0063487,

PubMed Abstract | Crossref Full Text | Google Scholar

Guzik, E. E., Byrge, C., and Gilde, C. (2023). The originality of machines: AI takes the Torrance test. J. Creat. 33:100065. doi: 10.1016/j.yjoc.2023.100065

Crossref Full Text | Google Scholar

Haase, J., and Hanel, P. H. P. (2023). Artificial muses: generative artificial intelligence chatbots have risen to human-level creativity. J. Creat. 33:100066. doi: 10.1016/j.yjoc.2023.100066

Crossref Full Text | Google Scholar

Habib, S., Vogel, T., Anli, X., and Thorne, E. (2024). How does generative artificial intelligence impact student creativity? J. Creat. 34:100072. doi: 10.1016/j.yjoc.2023.100072

Crossref Full Text | Google Scholar

Harvey, S., and Berry, J. (2022). Toward a meta-theory of creativity forms: how novelty and usefulness shape creativity. Acad. Manag. Rev. 48, 504–529. doi: 10.5465/amr.2020.0110

Crossref Full Text | Google Scholar

Hubert, K. F., Awa, K. N., and Zabelina, D. L. (2024). The current state of artificial intelligence generative language models is more creative than humans on divergent thinking tasks. Sci. Rep. 14:3440. doi: 10.1038/s41598-024-53303-w,

PubMed Abstract | Crossref Full Text | Google Scholar

Israilidis, J., Chen, W.-Y., and Tsakalerou, M. (2024). “Software development and education: transitioning towards AI enhanced teaching” in Proceedings of the 2024 IEEE global engineering education conference (EDUCON), Kos Island, Greece.

Google Scholar

Ivcevic, Z., and Grandinetti, M. (2024). Artificial intelligence as a tool for creativity. J. Creat. 34:100079. doi: 10.1016/j.yjoc.2024.100079

Crossref Full Text | Google Scholar

Kirkpatrick, K. (2023). Can AI demonstrate creativity? Commun. ACM 66, 21–23. doi: 10.1145/3575665

Crossref Full Text | Google Scholar

Koivisto, M., and Grassini, S. (2023). Best humans still outperform artificial intelligence in a creative divergent thinking task. Sci. Rep. 13:13601. doi: 10.1038/s41598-023-40858-3,

PubMed Abstract | Crossref Full Text | Google Scholar

Lee, B. C., and Chung, J. (2024). An empirical investigation of the impact of ChatGPT on creativity. Nat. Hum. Behav. 8, 1906–1914. doi: 10.1038/s41562-024-01953-1,

PubMed Abstract | Crossref Full Text | Google Scholar

Lockhart, E. N. S. (2025). Creativity in the age of AI: the human condition and the limits of machine generation. J. Cult. Cogn. Sci. 9, 83–88. doi: 10.1007/s41809-024-00158-2

Crossref Full Text | Google Scholar

Lukhmanov, Y., Perveen, A., and Tsakalerou, M. (2025). “ChatGPT in engineering teaching & learning: student and faculty perspective” in Proceedings of the 2025 IEEE global engineering education conference (EDUCON), London, United Kingdom.

Google Scholar

Marrone, R., Cropley, D., and Medeiros, K. (2024). How does narrow AI impact human creativity? Creat. Res. J. 1–11. doi: 10.1080/10400419.2024.2378264

Crossref Full Text | Google Scholar

Marrone, R., Taddeo, V., and Hill, G. (2022). Creativity and artificial intelligence—a student perspective. J. Intelligence 10:65. doi: 10.3390/jintelligence10030065,

PubMed Abstract | Crossref Full Text | Google Scholar

Oakley, B., Johnston, M., Chen, K., Jung, E., and Sejnowski, T. (2025). “The memory paradox: why our brains need knowledge in an age of AI” in The artificial intelligence revolution: challenges and opportunities. eds. M. Rangeley and L. N. Fairfax (Springer Nature: in press).

Google Scholar

Runco, M. A. (2023). AI can only produce artificial creativity. J. Creat. 33:100063. doi: 10.1016/j.yjoc.2023.100063

Crossref Full Text | Google Scholar

Seelig, T. (2017). Creativity rules (reprint). New York: HarperOne.

Google Scholar

Stevenson, C., Smal, I., Baas, M., Grasman, R., and Maas, H. (2022). Putting GPT-3’s creativity to the (alternative uses) test. In Proceedings of the 13th International Conference on Computational Creativity: ICCC’22. eds M. M. Hedblom, A. A. Kantosalo, R. Confalonieri, O. Kutz, and T. Veale (Bozen-Bolzano), 164–168. doi: 10.48550/arXiv.2206.08932

Crossref Full Text | Google Scholar

Summers-Stay, D., Lukin, S. M., and Voss, C. R. (2023). Brainstorm, then select: a generative language model improves its creativity score. Proceedings of the AAAI-23 Workshop Creative AI Across Modalities, Washington, DC, USA. Available at: https://api.semanticscholar.org/CorpusID:259305709

Google Scholar

Torrance, E. P. (1966). Torrance tests of creative thinking—norms technical manual research edition—verbal tests, forms A and B—figural tests, forms A and B. Princeton: Personnel Pres. Inc.

Google Scholar

Vinchon, F., Lubart, T., Bartolotta, S., Gironnay, V., Botella, M., Bourgeois-Bougrine, S., et al. (2023). Artificial intelligence & creativity: a manifesto for collaboration. J. Creat. Behav. 57, 472–484. doi: 10.1002/jocb.597

Crossref Full Text | Google Scholar

Wang, Y., and Lu, Y. (2025). Interaction, process, infrastructure: a unified architecture for human-agent collaboration. Available at: https://www.microsoft.com/en-us/research/publication/interaction-process-infrastructure-a-unified-framework-for-human-agent-collaboration/ (Accessed November 10, 2025).

Google Scholar

Wingström, R., Hautala, J., and Lundman, R. (2024). Redefining creativity in the era of AI? Perspectives of computer scientists and new media artists. Creat. Res. J. 36, 177–193. doi: 10.1080/10400419.2022.2107850

Crossref Full Text | Google Scholar

Keywords: artificial intelligence, creativity, design synthesis, engineering education, human–AI collaboration

Citation: Tsakalerou M, Akhmadi S, Balgynbayeva A and Kumisbek Y (2026) AI-assisted design synthesis and human creativity in engineering education. Front. Artif. Intell. 9:1714523. doi: 10.3389/frai.2026.1714523

Received: 27 September 2025; Revised: 23 December 2025; Accepted: 05 January 2026;
Published: 20 January 2026.

Edited by:

Sukhjit Singh Sehra, Wilfrid Laurier University, Canada

Reviewed by:

Paweł Wawrzała, Silesian University of Technology, Poland
Ezra N. S. Lockhart, Northcentral University, United States

Copyright © 2026 Tsakalerou, Akhmadi, Balgynbayeva and Kumisbek. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Saltanat Akhmadi, c3NhbHBlbm92YUBudS5lZHUua3o=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.