Your new experience awaits. Try the new design now and help us make it even better

METHODS article

Front. Artif. Intell., 16 September 2025

Sec. Medicine and Public Health

Volume 8 - 2025 | https://doi.org/10.3389/frai.2025.1644084

This article is part of the Research TopicGenAI in Healthcare: Technologies, Applications and EvaluationView all 9 articles

Privacy-, linguistic-, and information-preserving synthesis of clinical documentation through generative agents

Mark van Velzen,&#x;Mark van Velzen1,2Robert F. van der Willigen,,
&#x;Robert F. van der Willigen1,3,4*Vincent J. de Beer,Vincent J. de Beer1,3Helen I. de Graaf-Waar,Helen I. de Graaf-Waar1,2Esther R. C. Janssen,,Esther R. C. Janssen5,6,7Sjemaine van LeeuwenSjemaine van Leeuwen8Micha F. van der Willigen,Micha F. van der Willigen1,4Martijn J. van der Willigen,,Martijn J. van der Willigen1,3,4Gavin Renardus,Gavin Renardus1,4Rayan El Maaroufi,Rayan El Maaroufi1,4Sven J. Satimin,Sven J. Satimin1,4Larissa M. HartogLarissa M. Hartog1Tim Hulsen,Tim Hulsen1,9Nico L. U. van Meeteren,Nico L. U. van Meeteren2,10Mark C. Scheper,,,Mark C. Scheper1,4,11,12
  • 1Data Supported Healthcare: Data-Science Unit, Research Center Innovations in Care, Rotterdam University of Applied Sciences, Rotterdam, Netherlands
  • 2Department of Anesthesiology and Department of Cariothoracic Surgery, Erasmus Medical Center, Rotterdam, Netherlands
  • 3HR Datalab EAS, School of Engineering and Applied Science, Rotterdam University of Applied Sciences, Rotterdam, Netherlands,
  • 4School of Communication, Media and Information Technology, Rotterdam University of Applied Sciences, Rotterdam, Netherlands
  • 5Department of Orthopedic Surgery, VieCuri Medical Centre, Venlo, Netherlands
  • 6Radboud Institute for Health Sciences, IQ Health, Radboud University Medical Center, Nijmegen, Netherlands
  • 7School of Allied Health, HAN University of Applied Sciences, Nijmegen, Netherlands
  • 8Medifit Bewegingscentrum, Oss, Netherlands
  • 9Data Science & AI Engineering, Philips, Eindhoven, Netherlands
  • 10Top Sector Life Sciences and Health (Health~Holland), The Hague, Netherlands
  • 11Allied Health Professions, Faculty of Medicine and Science, Macquarrie University, Sydney, NSW, Australia
  • 12Solid Start Coalition, Erasmus Medical Center, Rotterdam, Netherlands

The widespread adoption of generative agents (GAs) is reshaping the healthcare landscape. Nonetheless, broad utilization is impeded by restricted access to high-quality, interoperable clinical documentation from electronic health records (EHRs) due to persistent legal, ethical, and technical barriers. Synthetic health data generation (SHDG), leveraging pre-trained large language models (LLMs) instantiated as GAs, could offer a practical solution by creating synthetic patient information that mimics genuine EHRs. The use of LLMs, however, is not without issues; significant concerns remain regarding privacy, potential bias propagation, the risk of generating inaccurate or misleading content, and the lack of transparency in how these models make decisions. We therefore propose a privacy-, linguistic-, and information-preserving SHDG protocol that employs multiple context-aware, role-specific GAs. Guided by targeted prompting and authentic EHRs—serving as structural and linguistic templates—role-specific GAs can, in principle, operate collaboratively through multi-turn interactions. We theorized that utilizing GAs in this fashion permits LLMs not only to produce synthetic EHRs that are accurate, consistent, and contextually appropriate, but also to expose the underlying decision-making process. To test this hypothesis, we developed a no-code GA-driven SHDG workflow as a proof of concept, which was implemented within a predefined, multi-layered data science infrastructure (DSI) stack—an integrated ensemble of software and hardware designed to support rapid prototyping and deployment. The DSI stack streamlines implementation for healthcare professionals, improving accessibility, usability, and cybersecurity. To deploy and validate GA-assisted workflows, we implemented a fully automated SHDG evaluation framework—co-developed with GenAI technology—which holistically compares the informational and linguistic features of synthetic, anonymized, and real EHRs at both the document and corpus levels. Our findings highlight that SHDG implemented through GAs offers a scalable, transparent, and reproducible methodology for unlocking the potential of clinical documentation to drive innovation, accelerate research, and advance the development of learning health systems. The source code, synthetic datasets, toolchains and prompts created for this study can be accessed at the GitHub repository: https://github.com/HR-DataLab-Healthcare/RESEARCH_SUPPORT/tree/main/PROJECTS/Generative_Agent_based_Data-Synthesis.

1 Introduction

While the use of GenAI in healthcare offers substantial promise and is expected to become an integral part of regular clinical practice, its widespread adoption is limited by several critical challenges (Sai et al., 2024). These include fragmented and often inaccessible data silos, significant variability in the quality of real-world data, and complex ethical and legal considerations. Efforts to address these barriers are underway, including the implementation of federated learning approaches, the establishment of robust data governance frameworks that incorporate standards such as HL7 FHIR and FAIR principles, and the development of evolving regulatory measures—such as the European Union AI Act and the General Data Protection Regulation. Collectively, these initiatives aim to foster a landscape of responsible and ethical innovation in healthcare AI (Mons et al., 2017; Busch et al., 2024; Woisetschläger et al., 2024; Ibrahim et al., 2025; Liu et al., 2025). For readers seeking concrete real-world clinical practice use cases of GenAI in healthcare information systems, see Sai et al. (2024) and Reddy (2024).

Real-world data as recorded in EHRs consists primarily of free-text narratives that often contain valuable clinical insights that are not captured by structured data formats (Negro-Calduch et al., 2021; Seinen et al., 2024). However, the synthesis of meaningful health-related text has received little attention in GenAI literature (Murtaza et al., 2023; Ibrahim et al., 2025; Loni et al., 2025; Rujas et al., 2025). This represents a missed opportunity, not only in terms of data synthesis methodology, but also because free-text narratives often contain hidden and nuanced clinical information that is essential for meaningful patient understanding, accelerated early diagnosis, and improved clinical decision making (Negro-Calduch et al., 2021; Seinen et al., 2024; Schut et al., 2025). Moreover, while methods for synthesizing structured EHR data are well established, the generation of synthetic free text from real-world EHRs remains a comparatively underdeveloped area (Murtaza et al., 2023; Ibrahim et al., 2025; Rujas et al., 2025).

High quality synthetic health data generation (SHDG) that emulates real-world clinical documents is emerging as a vital strategy to enable safe, scalable, and privacy-preserving GenAI implementation in health and care settings (Smolyak et al., 2024; Loni et al., 2025). To be precise, SHDG is a privacy-enhancing technology that entails the generation of synthetic data based on real-world datasets. These synthetic datasets are designed to retain the essential statistical patterns and relationships of the original data, without containing any directly identifiable information. The goal is to enable analyses on synthetic data that produce results closely mirroring those obtained from the real data (Murtaza et al., 2023; Drechsler and Haensch, 2024; Ibrahim et al., 2025; Rujas et al., 2025). Ideally fully interoperable and machine-readable, synthetic datasets provide a foundation for the development, testing, and validation of innovative applications ranging from personalized healthcare models to solutions that alleviate administrative workload (Ibrahim et al., 2025; Rujas et al., 2025).

The growing promise and pitfalls of SHDG is best understood in the context of major breakthroughs in natural language processing (NLP) and GenAI, spanning early statistical approaches to today’s transformer-based language models. Synthetic data generation gained momentum in the early 1990s exemplified by Rubin’s (Rubin, 1993) multiple imputation framework and Little’s efforts on statistical disclosure through data masking (Little, 1993). Since 2010 the adoption of machine learning and deep learning has expanded the applications of synthetic data, especially in healthcare (Eigenschink et al., 2023; Murtaza et al., 2023; Drechsler and Haensch, 2024; Goyal and Mahmoud, 2024; Pezoulas et al., 2024). Deep learning models can automatically extract and represent intricate patterns from large datasets, thereby facilitating more accurate and efficient processing of information. Early approaches used models called convolutional and recurrent neural networks, which allowed machines to automatically process unstructured datasets such as images and speech without the need for feature analysis. However, deep learning model development is labor-intensive and complex (LeCun et al., 2015; Goodfellow et al., 2018).

The use of Generative Adversarial Networks (GANs) has led to significant advances in data synthesis and can be classified as both deep learning and GenAI approaches. GANs rely on adversarial training, in which two deep neural networks—the generator and the discriminator—compete with one another until the discriminator can no longer distinguish between real and synthetic data, reaching equilibrium (Goodfellow et al., 2018; Baowaly et al., 2019; Ibrahim et al., 2025). It is the most widely adopted AI technology in the fields of health and healthcare for generating synthetic data. GAN models were originally developed to produce realistic pictures—such as images from magnetic resonance imaging (MRI) or dermatoscopy—as well as to create continuous data. Continuous data includes numerical values that can take on any value within a range, like blood pressure, glucose levels, or patient age. In addition, specialized GAN models have also been used to generate time-series data, such as signals from electrocardiograms (ECG) and electroencephalograms (EEG) (Ibrahim et al., 2025).

The advent of transformer-based architectures revolutionized NLP by significantly enhancing language understanding (Vaswani et al., 2017). Groundbreaking pre-trained language models such as “bidirectional encoder representations from transformer” (BERT) laid the groundwork for this transition by introducing mechanisms for capturing nuanced contextual relationships within text. Building upon this foundation, domain-specific adaptations like BioBERT and Clinical-BERT, trained on biomedical literature and clinical notes, respectively, substantially increased the fidelity with which deep learning could represent the intricate semantics of medical language through vectorization and embeddings of text into numerical expressions (Alsentzer et al., 2019; Lee et al., 2020; Bommasani et al., 2021; Ibrahim et al., 2025). Capturing clinically meaningful concepts through tokenization, vectorization, and text embedding, however, remains a formidable challenge. For example, clinical notes contain valuable contextual information but are characterized by a variety of nomenclatures, abbreviations, misspellings, and synonyms both within and across healthcare disciplines (Cannon and Lucci, 2010; Doan et al., 2014; Meystre et al., 2017; Negro-Calduch et al., 2021; Seinen et al., 2024; Schut et al., 2025).

The evolution of transformer-based architectures culminated in the development of Large Language Models (LLMs) that are not only capable of language understanding—like BERT—but are also capable of language generation—like OpenAI’s GPT3—, called foundational models (Bommasani et al., 2021). LLMs further amplified generative and NLP capabilities—made possible by scaling up model architectures, leveraging enhanced computational power, and training on extremely large datasets. The emergence of novel functionality is associated with the so-called “scaling effect”—a phenomenon that was initially unforeseen and, to this day, remains inadequately understood (Kaplan et al., 2020; Coveney and Succi, 2025). Recent comprehensive surveys underscore the complexity and ongoing debate surrounding the generality of scaling laws, highlighting both their impressive predictive successes and significant limitations—especially when LLMs are applied in situations quite different from the functionality they were originally trained for, or when they face new types of data they have not seen before (Li et al., 2023; Sengupta et al., 2025). One of the most striking emergent properties of LLMs is their ability to respond to prompting—where users provide specific instructions or examples in natural language to guide the model’s output. This phenomenon allows for highly flexible and tailored use of LLMs without the need for additional fine-tuning, thereby broadening their practical utility in clinical document generation and comprehension. For example, prompting allows for zero-shot and few-shot learning, wherein LLMs can generate appropriate responses to tasks or questions with either no examples (zero-shot) or just a handful of provided examples (few-shot), greatly enhancing their versatility to synthesize realistic clinical narratives (Brown et al., 2020).

The rapid development of efficient, fine-tunable Small Language Models (SLMs) and multimodal SLMs is rapidly transforming healthcare. Unlike their bigger LLM counterparts SLMs can operate on-premises servers supporting local deployment, which significantly reduces costs, carbon footprint, and privacy concerns. These advances—as well as techniques like quantization, which reduce model size and accelerate inference time—are making it feasible to bring powerful language understanding and generation capabilities directly to the point of care, enabling clinical decision support, documentation, and patient interaction without reliance on high-end cloud infrastructure (Schick and Schütze, 2021; Dibia et al., 2024; Garg et al., 2025; Kim et al., 2025; Xie et al., 2025).

However, as GenAI becomes increasingly integrated into clinical workflows, it faces unique challenges specific to the medical domain. One of the most persistent and technically demanding issues is the precise segmentation and representation of multi-word clinical terms—such as “low blood pressure” or “low back pain” as well as their common abbreviations like “LBP.” Language models, including both LLMs and SLMs, often struggle to consistently recognize and encode such terms as unified clinical concepts. Inconsistent tokenization or lack of domain-specific context can lead to fragmented or distorted semantic representations, which in turn may compromise the accuracy of clinical information extraction, decision support, or narrative synthesis. This underscores the ongoing need for the development of more sophisticated prompting strategies tailored for healthcare, and the integration of concept embeddings, clinical practice guidelines, and real-world sample data to improve the realism and utility of synthetic clinical data (Beam et al., 2020; Chung et al., 2023; Han et al., 2023; Li et al., 2023).

2 GA-assisted SHDG workflows

To overcome the challenges of semantic fragmentation and context loss, recent research has shifted toward collaborative, multi-agent workflows to jointly tackle the complexity of clinical narratives and multimodal data. This new class of GenAI—known as Agentic AI or Generative Agents (GAs)—features multiple autonomous agents, each specializing in a particular task or sensory domain such as vision, language, audio, or touch (Chan et al., 2023; Park et al., 2023). That is, each agent can preside over its own dedicated LLM, prompted to address the unique challenges of its modality and dedicated task. By collaborating iteratively and sharing insights across these different modalities, GAs are able to process and integrate information from multiple sources—text, images, speech—generating content that is coherent and contextually rich (Qiu et al., 2024; Gridach et al., 2025; Hettiarachchi, 2025; Piccialli et al., 2025; Schneider, 2025).

Unlike traditional GenAI, GAs can gather real-time data from various sources, use different tools, design custom workflows, and refine their strategies through feedback, making them highly flexible and context aware. This collaborative approach allows specialized agents to solve complex tasks—such as SHDG—within a single workflow (Qiu et al., 2024). Some advanced workflows also incorporate language-agnostic concept models, which can further enhance the quality and performance of data synthesis (Daull et al., 2023; Barrault et al., 2024; Xie et al., 2024; Jin et al., 2025).

Building on this flexible and context-aware foundation, role-specific GAs are able to distribute cognitive workloads and leverage the strengths of multiple specialized agents within a single workflow (Xie et al., 2024; Jin et al., 2025). This not only allows for more effective problem solving but also enables the application of domain-specific expertise and enhances system robustness across a range of complex real-world tasks. For example, by leveraging advanced language models such as GPT-4, GAs can support workflows in scientific literature review writing (Abdurahman et al., 2025), sentiment analysis (Vasireddy et al., 2024), and the identification of mental health symptoms such as social anxiety from clinical interview data (Ohse et al., 2024). GAs also demonstrate promise in interpreting and structuring unstructured data from EHRs and clinical documentation (Yang et al., 2025). In particular, GPT-4.1 can complement—and sometimes rival—human expertise in global health education and data analysis (Thandla et al., 2024).

Low-code/no-code LLM platforms, such as Flowise,1 Langflow,2 and AutoGen3 (Kumar, 2023), are transforming the way GA-assisted SHDG workflows could be developed (Jeong, 2025). These platforms use drag-and-drop interfaces and natural language prompts, making it much easier to create, modify, and optimize workflows that leverage LLMs without the requirement of in-depth coding expertise. As a result, healthcare professionals and other non-technical users can now play an active role in designing and improving intelligent systems relevant to their needs. Moreover, a recent development called “Vibe-coding”—coined by Karpathy (2025)—illustrates this shift by enabling conversational co-development. Here, developers interact with GenAI tools—like GitHub Copilot—using plain language to iteratively refine and adjust code (Karpathy, 2025; Mayo, 2025).

In what follows, we present a protocolized GA-driven proof-of-concept for SHDG use cases aimed at creating novel synthetic clinical documents that emulate genuine EHRs while safeguarding patient privacy, preserving linguistic integrity, and maintaining informational accuracy under conditions of limited access to authentic EHR datasets.

3 Materials and methods

The protocol described here aims to expand the responsible use and deployment of GenAI healthcare solutions to a broad spectrum of end-users—including those without specialized AI expertise—by providing clear, step-by-step guidance for designing workflows that leverage GAs for SHDG. To further support open-source adoption, reproducibility and practical application of our GA-assisted SHDG workflows we provide a GitHub repository.4

At the heart of our methodology is a modular data science infrastructure (DSI) Stack (Figure 1A), designed to organize and streamline the process of generating clinical data by breaking it down into a series of well-defined, interconnected workflows. It starts with ingesting and converting anonymized clinical notes from PDF to Markdown (FLOW01) (see text footnote 4), followed by pseudonymization to protect patient privacy (FLOW02) (see text footnote 4). Synthetic notes are then produced using LLMs, comparing standard prompting (FLOW03) (see text footnote 4) with a Generative Agent approach (FLOW03_AGENT_BASED). The pipeline ends with benchmarking (FLOW04) (see text footnote 4) using metrics for diversity, vocabulary similarity, semantic alignment, and classifier-based machine discernibility—enabling the comparison of the informational and linguistic features of synthetic, anonymized, and real EHRs at both the document and corpus levels. Included is a hands-on guide for responsible deployment of GA-assisted SHDG-workflows,5 implemented through open-source data science platforms—Hugging Face Spaces and Flowise—and powered by LLMs accessed via public cloud services such as Azure. This hybrid approach enables rapid prototyping and controlled sharing of workflows, while API key–secured inference endpoints ensure privacy compliance. The GitHub repository also provides lines of example source code (FLOW03) (see text footnote 4) that enables implementation of a privacy-first alternative to public cloud AI services called Ollama, allowing users to run and manage large language models locally while maintaining full data control and privacy compliance.

Figure 1
Diagram comparing data science and data engineering roles. Panel A shows stages from warehousing to deployment, divided into data science and engineering tasks. Data science involves deployment, feature engineering, and model development. Data engineering includes warehousing, compute, and toolchain tasks. Coding infrastructure is central. Data science focuses on doing things right, while data engineering focuses on doing the right things. Panel B contrasts human-oriented data scientists with machine-oriented data engineers, emphasizing their differing involvement levels.

Figure 1. Data science infrastructure (DSI) stack. (A) Schematic of the DSI stack, structured as modular, interoperable layers founded on key IT principles: abstraction and modularization, separation of concerns, interoperability and standardization, scalability, and resilience. Each layer—[1] … [8]—fulfils a distinct function, from data storage and processing to analytics and deployment. It supports flexible, maintainable, and scalable data science pipelines. The DSI stack aligns with the Double Diamond design model (https://www.designcouncil.org.uk/our-resources/the-double-diamond/). Lower layers focus on “Doing the right things”—data gathering and integration—, while upper layers emphasize “Doing things right”—curation, iteration, deployment. (B) Visualization of the roles and involvement of data scientists versus data engineers across the DSI stack. While data scientists are predominantly active in the human-oriented, upper layers of feature engineering, model development, and deployment, data engineers are primarily engaged in the machine-oriented, foundational layers involving warehousing, compute, and toolchains. This panel highlights the complementary skill sets necessary for an effective and robust data science infrastructure.

3.1 Data science infrastructure stack

Synthesizing and validating EHRs requires a well-defined data science infrastructure. This involves designing a robust data pipeline, a systematic sequence of processes that transforms raw data into high-quality synthetic datasets and generates actionable insights (Meng, 2021; Tuulos, 2022).

Building a pipeline requires the seamless integration of diverse hardware and software components within a pre-defined DSI stack. In this architecture, each layer—from data ingestion to deployment—builds upon the previous one, creating a cohesive framework. This layered approach streamlines the transformation of real-world EHR samples and clinical practice guidelines into novel, synthesized datasets. By following this structured process, patient privacy and regulatory compliance are maintained, enabling the safe and effective use of synthetic data for research, analytics, and clinical decision support (Priebe et al., 2021; Tuulos, 2022; Hechler et al., 2023).

Our DSI stack—as depicted in Figure 1—is organized into eight layers: [1] Data Warehousing; [2] Compute Resources; [3] Toolchain; [4] Workflow Orchestration; [5] Software Architecture; [6] Model Development; [7] Feature Engineering; [8] Data Product Deployment. The dependency on IT-hardware progressively increases towards the bottom of the stack (Krishnakumar et al., 2023). Next, we describe each of the relevant layers from the bottom up, highlighting their importance and practical application. Note that layers [6] Model Development and [7] Feature Engineering were not applicable to our specific use case and will therefore not be discussed further. Moreover, our DSI stack aligns with the double diamond design model (Kochanowska et al., 2022): lower layers focus on “Doing the right things,” while upper layers emphasize “Doing things right.”

3.1.1 Pseudonymization

Integrating real-world sample data into the synthesization process enhances the diversity of synthetic data, thereby improving its linguistic quality and informational characteristics to better reflect real-world EHR narratives (Chung et al., 2023). Clinical narratives were collected from 13 patients undergoing treatment for low back pain at a single physiotherapy clinic within a primary care setting, spanning the duration of their therapeutic trajectories. All patients provided informed consent for the use of their EHRs to assist in the generation of synthetic data. All EHRs (N = 13) were manually anonymized by deleting patient names, addresses, social security numbers, contact details, and insurance details.

Pseudonymization served as an essential data pre-processing step prior to warehousing (Section 3.1.2). This involved replacing the names of referring physicians and treating physiotherapists with fictive names, thereby restoring the natural structure of the EHRs. This procedure ensured that sample datasets destined for warehousing were thoroughly de-identified in accordance with Dutch and broader European privacy and regulatory standards. To achieve this, we developed a GenAI-based Named Entity Recognition (NER) workflow, customized for privacy categories, to systematically identify and replace personal identifiers in Markdown files derived from EHR sample PDF documents. Data entry fields for entities such as names, addresses, contact details, birth dates, “burgerservicenummers” (BSNs), insurance details, and financial data were detected and either removed or pseudonymized in compliance with privacy guidelines.

A Jupyter notebook executed custom Python code [GitHub Repository (see text footnote 4): FLOW01] to configure the Azure OpenAI API SDK; submit each document to Azure OpenAI’s GPT-4.1; and apply a tailored system prompt to both identify and pseudonymize specified entities while preserving the Markdown format. Processed files were stored with their original formatting intact for subsequent analysis.

3.1.2 Warehousing

Data warehousing serves as the foundational layer of the DSI stack, providing centralized aggregation and accessibility for static, unstructured datasets. It is crucial for storing both the generated synthetic data (output for developing and testing solutions) and the real-world sample data and clinical practice guidelines (as input knowledge bases for the synthesis process). Besides, the integration of anonymized real-world data from EHR systems, the inclusion of codebooks for labels and abbreviations, and clinical practice guidelines into data warehouses limits chance of hallucination but enhances the diversity, linguistic quality, and therefore the clinical relevance of synthetic data (Chung et al., 2023; Li et al., 2023).

Storing data in accessible formats such as markdown (MD), structured query language (SQL), comma-separated values (CSV), portable document format (PDF), JavaScript object notation (JSON), or images is integral to facilitating interoperability and data sharing within clinical environments. These widely used formats support the storage and exchange of both structured and unstructured data, making it easier for diverse health information systems to work together (Hart et al., 2016). For example, JSON is used in AI-workflows because it is both human-readable and machine-readable. It is particularly helpful in health care and other fields for quickly and efficiently transferring data between systems, applications, or devices.

3.1.3 Compute

The next DSI stack layer is compute, which refers to scalable data processing capacity or computational power. Its purpose is to manage and scale the performance of a predefined set of computational instructions—referred to as a computational workload or task. The type of data science use case dictates its compute requirements, including the need for CPUs, GPUs, TPUs, or internal memory.

To assist the targeted end-users—non-AI specialists—, we decided to employ a hybrid compute solution, combining standard desktop computers or laptops (local computing) with powerful external computing resources available over the internet (public cloud services like Azure, AWS, or Google Cloud). The latter is essential for utilizing state-of-the-art LLMs. These models require specialized high-performance computing hardware and massive computational resources that far exceed the capabilities of standard desktops or laptops.

Public cloud enables scalable LLM deployment with on-demand access to high-end GPUs/TPUs, large memory, and parallel computing, surpassing local infrastructure limits. Since state-of-the-art LLMs are only accessible via cloud-based APIs (Table 1), cloud adoption is essential for high-end performance. However, organizations with sufficient local computing power (e.g., EHR vendors, hospitals) may opt for on-premises GenAI model deployment.

Table 1
www.frontiersin.org

Table 1. Key evaluation criteria for selecting LLMs.

3.1.4 Toolchain

The toolchain layer ensures the correct functioning of the desired workflow orchestration. Our protocol leverages rapid prototyping platforms to synthesize and validate EHRs using state-of-the-art GA-assisted SHDG workflows. These platforms offer an intuitive drag-and-drop user interface, simplifying implementation by eliminating the need for data engineering expertise in LLM deployment. This makes GenAI-technology more accessible for non-AI specialists by facilitating browser-based access (Kumar, 2023; Jeong, 2025).

We implemented a Docker-based6 architecture to enable secure, reproducible workflows across on-premises and public cloud environments. By deploying containerized workflows via API key–secured inference endpoints, we ensure scalable resource allocation, consistent performance, and robust access control (Abhishek and Rao, 2021; Ait et al., 2025). Inference endpoints offer a user-friendly interface for submitting inputs—such as prompts for synthetic EHR generation—and receiving outputs, allowing on-demand use of custom GA-assisted workflows without exposing users to underlying infrastructure or model complexity (Fu et al., 2025; Gupta, 2025).

Hugging Face Spaces7 offers a public cloud platform for rapidly building, sharing, and interacting with containerized GenAI applications through intuitive interfaces, automatically hosting workflows at a public URL. This supports Docker-based deployment of low-code/no-code tools such as Flowise, Langflow and AutoGen for developing multi-agent AI workflows (Kumar, 2023; Jeong, 2025). Flowise enables users to visually compose, configure, and deploy LLMs without programming expertise, while Langflow allows for direct code customization. AutoGen Studio facilitates rapid prototyping and orchestration of LLM-based multi-agent systems via a low-code Python framework. For example, an implementation guide titled “Learn how to deploy Flowise on Hugging Face” is available online.8

The usefulness of Open-source LLMs like LLaMA, Qwen, DeepSeek, and Phi in clinical settings is hampered by the frequent production of unsupported facts, contradictions, and omissions—collectively known as hallucinations—which present a substantial safety risk. Evaluation of the MIMIC-IV dataset revealed that while these models can accurately capture up to 83% of admission reasons and key events, their performance dropped dramatically for critical follow-up recommendations, with comprehensive coverage as low as 29% (Das et al., 2025).

Considering both the above outlined limitations and the selection criteria presented in Table 1, we selected OpenAI’s GPT-4.1 (version:2025-04-14) as our preferred LLM. GPT-4.1 offers a one million token context window, which allows for comprehensive analysis of extensive patient records, and demonstrates robust instruction-following and advanced reasoning abilities (OpenAI, 2025). The criteria listed in Table 1 reflect insights drawn from recent expert reviews and leading practitioner models (QuantSpark, 2023; Inoue, 2024; Ruczynski, 2024; Chojnacki, 2025; Morris et al., 2025).

In addition, GPT-4.1 incorporates advanced domain adaptation, sophisticated fact-checking mechanisms, and alignment strategies to reduce hallucination rates and enhance faithful adherence to source texts. Here, “Advanced domain adaptation” refers to the ability of an LLM to adjust its understanding and generation of text to fit the specific language, conventions, and knowledge of a particular field or domain—such as medicine (Singhal et al., 2023). Moreover, GPT-4.1 appears to demonstrate advanced comprehension of medical and healthcare language, enabling more accurate interpretation of complex EHR narratives and improved detection of inconsistencies (Walturn, 2025).

3.1.5 Workflow orchestration

To synthesize Dutch clinical narratives related to low back pain in physiotherapy using genuine EHRs, we developed a GA-assisted, no-code workflow built on a rapid prototyping platform that allows end users to quickly construct and test GenAI solutions (Figure 2). This platform is organized into modular components—known as modules—each serving a specific purpose and designed to be easily rearranged or modified.

Figure 2
No-code multi-agent EHR synthesizing workflow diagram showing interconnected modules. Key components include Recursive Character Text Splitter, Pdf File input, In-Memory Vector Store, Retriever Tool, Azure OpenAI Embeddings, Azure ChatOpenAI, SQLite Agent Memory, Supervisor, and Worker. Each module has specific input and output parameters, facilitating the processing and synthesis of electronic health record data.

Figure 2. Visual representation of a no-code, multi-agent workflow for synthesizing EHRs. Data flows through connected tools and agents, enabling an iterative, structured generation process without manual coding. The here shown GA-assisted SHDG workflow begins with a Recursive Character Text Splitter that divides the uploaded PDF file containing anonymized EHR data into manageable chunks. These chunks are processed using Azure OpenAI Embeddings and stored in an In-Memory Vector Store. A Retriever Tool (RAG) then queries the stored embeddings to provide relevant context. The Azure ChatOpenAI component, configured with the GPT-4.0-mini model, interacts with stored agent memory (SQLite Agent Memory) and coordinates with two agents: the Supervisor—acting as a senior physiotherapist specialized in low back pain—who manages task instructions and workflow control, and directs the Tech Researcher—acting as a practicing physiotherapist (general or specialized—who executes prompts to generate synthetic Dutch-language EHR notes). Note: The workflow—accessed via a web interface—maintains contextual memory across user queries for seamless, multi-turn interactions and stops automatically when the supervisor determines completion. This design enables rapid prototyping of SHDG solutions by healthcare researchers and practitioners without requiring advanced expertise in AI. For more information on the adopted technologies and their implementation, see the Toolchain Section (Section 3.1.4 of the DSI stack).

Central to our approach is a multi-agent architecture (Figure 3), comprising a supervisor agent and one or more worker agents, each guided by tailored prompts that correspond to their unique responsibilities. To make the process intuitive, we utilized a “What-IF” scenario: imagine a clinical team with a general practitioner, a specialist, and a medical scribe collaborating to create a realistic synthetic patient note. In this analogy, the supervisor agent acts like the lead clinician—overseeing the entire process, distributing tasks, prioritizing activities, monitoring progress, and ensuring the workflow stays on track and determines when the task is finished—while the worker agents take on specialized roles such as drafting clinical content or formatting records, analogous to the scribe and specialist. Each agent receives role-specific guidance through prompt engineering (Chen et al., 2025a), ensuring adherence to clinical standards and the accurate, coherent assembly of narratives. This modular, role-based structure not only enables seamless coordination and iterative refinement among AI agents, but also produces synthetic EHRs with high clinical validity and practical value for healthcare workers.

Figure 3
Graphic displaying a structured layout for creating realistic physiotherapeutic patient files in Dutch, addressing low back pain. It includes sections on anamnesis, diagnosis following ICF, SMART therapeutic goals, and treatment plans. Text boxes offer guidelines for language use, workflow examples, and specify that 20 unique patient files are to be generated. The graphic is divided into roles: End-user, supervisor, and tech researcher, with a patient dossier example detailed. A status box at the bottom-right corner indicates

Figure 3. Example of a single-turn input/output interaction when applying the multi-agent workflow for generating realistic, structured EHRs, as described in Figure 2. In this scenario, the End User—a practicing physiotherapist—uses a web-based interface to request the creation of 20 synthetic but realistic EHRs in Dutch. The request specifies detailed content and formatting requirements, including: a concise patient history summary, a clearly stated help-seeking question, an ICF-based diagnosis, measurable treatment goals, and a treatment plan aligned with KNGF guidelines. All records must use professional Dutch clinical language with correct abbreviations. The Supervisor ensures these specifications are complete and unambiguous before the Tech Researcher produces a synthetic yet realistic Dutch-language EHR. Each record contains the requested summary, patient demographics, presenting complaint, ICF-based functional and contextual factors, SMART goals, an individualized treatment plan, and SOAP-formatted progress notes. For illustration purposes, only Patient Dossier 14 from the generated set is shown here. The Supervisor reviews this output, confirms it meets all requirements, and marks the task as finished. Color coding: Blue—End User; Pink—Supervisor Agent; Green—Tech Researcher (worker Agent).

3.1.5.1 Supervisor agent prompt

The supervisor agent was equipped with a “system prompt”—a foundational set of instructions that defines the LLM’s overarching persona, scope, and governance strategy. Specifically, this prompt directed the supervisor agent to represent an experienced physiotherapist overseeing the clinical perspective of a registered physiotherapist tasked with generating authentic Dutch EHRs for low back pain cases. The prompt specified how to generate clinical documentation according to the International Classification of Functioning, Disability and Health (ICF) framework domains (World Health Organization, 2001), and adherence to the Dutch Royal Society for Physiotherapy (KNGF) guideline on low back pain (Swart et al., 2021). This ensured that every generated record aligned with current best practices in physiotherapy documentation.

The prompt further instructed the supervisor agent to restrict outputs solely to the requested EHR content, thereby enforcing compliance and preventing the inclusion of extraneous or sensitive information. Functionally, it mandated the supervisor agent to orchestrate the division of labor among worker agents, manage task handoff, establish execution priorities, consolidate contributions, and verify completion of all record elements in accordance with professional documentation standards.

3.1.5.2 Worker agent prompt

Under the coordination of the supervisor agent, each worker agent received individualized “worker prompts” tailored to its specialized domain within the workflow. These prompts offered detailed, task-specific instructions for generating a single EHR taking into account the clinical nuances of physiotherapy care for (sub)acute or chronic low back pain, as well as relevant and documentation standards (Driehuis et al., 2019; Swart et al., 2021). Most notably, the worker prompts required structured output, as listed in Table 2.

Table 2
www.frontiersin.org

Table 2. Targeted worker agent prompting.

3.1.6 Software architecture and deployment

Data product deployment is the final layer of our DSI stack, where prototype workflows transform into web-accessible applications. This layer ensures that data products—such as LLM GPT-4.1 and Hugging Face Spaces—are securely and reliably made available through Inference Endpoints secured with API key authorization (as was discussed in the Toolchain Section 3.1.4). This also protects patient privacy and complies with data management and regulatory standards, including the GDPR9 and the EU AI Act10 (Haug, 2018; Hoofnagle et al., 2019; European Parliament and Council, 2024). Additionally, it manages resources and access controls to maintain data security and organization. By simplifying deployment through no-code/low-code platforms, discussed in the next section, with reusable components and clear specifications, this layer helps deliver trustworthy, easy-to-use solutions that support clinical decision-making, research, and analytics.

3.2 No-code proof-of-concept

This section details a rapid-prototyping implementation to demonstrate the feasibility and effectiveness of generating synthetic EHRs using GA-assisted SHDG workflows (Figures 2, 3). By combining modularity and no-code principles, our proof-of-concept illustrates a practical, user-friendly approach for healthcare professionals to generate, validate, and experiment with synthetic health data. It supports iterative development, transparency, and ease of integration with other systems or data science workflows.

Here, we outline the specific tools and configurations used, illustrating how the DSI stack layers translate a specific data science use case into a functional workflow. The protocol presented here serves as a practical step-by-step guide for replicating our approach and showcases the capabilities of the proposed architecture in addressing the challenges of clinical text synthesis.

For demonstrative purposes, we utilized public Hugging Face Spaces infrastructure in combination with Flowise to facilitate deployment (see Section 3.1.4). This setup allows custom-made workflows to be shared publicly or privately, making them accessible via a web interface or API, and supports secure credential management for connecting to external LLMs and API services.

The workflow (Figure 2) to synthesize EHRs, required additional modules for document ingestion, embedding generation (using Azure API key credentials), agent memory management (using a SQLite database) and memory retrieval through a vector store. Table 3 provides a functional overview of the Flowise modules shown in Figure 3. To digest unstructured sample data (e.g., PDFs), we applied a recursive character text splitter. Traditionally, parameters such as chunk size and chunk overlap are critical: smaller chunks (500–1,000 characters) improve granularity but may fragment context, while larger chunks (>1,500 characters) support coherence but risk exceeding model context windows. Moderate overlap (100–200 characters) helps maintain semantic continuity between segments. However, recent studies suggest that late chunking—where segmentation occurs after the model embedding step rather than before—can preserve global context more effectively and enhance downstream performance, particularly in retrieval-augmented tasks (Günther et al., 2024).

Table 3
www.frontiersin.org

Table 3. Workflow stages, modules, and operational details for no-code GA-assisted synthetic EHR processing.

We employed GPT-4.1 LLMs—using Azure API key credentials—for supervised reasoning and text generation. A key model parameter for any LLM is temperature, which controls generative diversity: lower values (~0.2–0.4) yield deterministic, guideline-conform output, while higher values (~0.7–0.9) promote creative variability (Peeperkorn et al., 2024). For clinical synthesis, we assumed that a temperature between 0.3–0.5 best balances realism and consistency; however, future experiments are required to empirically evaluate model performance at varying settings. The “agentflow” as implemented within the Flowise no-code framework can be downloaded as a JSON file from our GitHub Repository (see text footnote 1) (GA-assisted SHDG workflow).

The supervisor agent represents an experienced physiotherapist who ensures that documentation conforms to ICF domains, linguistic plausibility, and overall fidelity (Figure 3). The worker agent (Figure 3) embodies one of several physiotherapy profiles (e.g., generalist, manual therapist, exercise therapist, psychosomatic physiotherapist) and is tasked with generating synthetic clinical narratives. To reduce the risk of hallucination and ensure domain-conformant output, we implemented a retrieval-augmented generation (RAG) pipeline. This supported the contextual grounding of worker agent output using a vectorized memory store filled with real-world sample data, clinical practice guidelines, and documentation standards used by Dutch physiotherapists. This architecture allows agents to generate clinically realistic output grounded in both empirical input and normative context (Chung et al., 2023; Li et al., 2023).

3.3 SHDG automation through GenAI-assisted co-development

Here we describe how we automatized the entire synthetic EHR pipeline; using a novel engineering approach called: GenAI-assisted co-development. This technique fosters collaboration between human developers and GAs (Park et al., 2023) allowing human domain specialists to iteratively refine code through natural language prompts and AI-tools like GitHub Copilot, Perplexity, and Gemini. Copilot helps with coding by suggesting and writing code, Perplexity helps find answers by providing clear, sourced information, and Gemini acts as an all-around smart assistant that can respond to questions, summarize text, and assist with a variety of tasks. As such, GenAI-assisted co-development fundamentally redefines the paradigm of human-AI collaboration (Orru et al., 2023; Ulfsnes et al., 2024; Casper et al., 2025; Mayo, 2025).

For instance, GAs like AlphaEvolve (DeepMind, 2025) exemplify the power of GenAI-assisted co-development by autonomously facilitating iterative code improvement. AlphaEvolve operates as an evolutionary coding agent that leverages the orchestration of multiple LLMs, enabling ongoing refinement of algorithmic solutions through a cycle of edits and evaluator feedback. This process mirrors the collaborative workflow described earlier (Section 3.2), whereby human expertise and GAs interleave in a continuous dialogue—using natural language guidance to steer, critique, and optimize computational problem-solving. In this way, AlphaEvolve and similar systems advance the core objective of our methodology: to solve complex scientific and computational challenges by seamlessly integrating human insight with autonomous generative capabilities (Novikov et al., 2025).

3.3.1 GA-assisted SHDG workflow validation

We started by confirming that our GA-assisted SHDG workflow (Section 3.2) was effective in generating meaningful synthetic EHRs. This preliminary validation step involved human specialists (authors MV, MS) assessing the synthetic EHR samples for realism, internal coherence, and adherence to professional clinical documentation standards. We confirmed that the generated records were representative for deployment in downstream healthcare applications and research.

3.3.2 PDF-to-Markdown conversion

As a foundational step, we used Gemini 2.5 Flash as the core LLM in our GenAI-Assisted Software Development workflow to generate and refine Python code—contained in Jupyter notebooks—for automating PDF-to-Markdown conversion. Guided by natural language prompts written by human domain experts, Gemini 2.5 Flash authored code that extracts text from PDFs (via Python packages such as OpenAI, PyMuPDF, and glob), interfaces with Azure OpenAI’s GPT-4.1 for Markdown formatting, and supports batch processing of files. This collaborative approach enabled efficient, maintainable code development, combining Gemini 2.5 Flash’s reasoning and coding abilities with GPT-4.1’s language understanding to deliver scalable and accurate document conversion. Note, Human supervision— following a human-in-the-loop approach, in which humans remain actively involved in reviewing, verifying, and refining AI outputs—was essential to ensure that the AI-generated code functioned correctly. The code used is available online via our GitHub Repository (see text footnote 1) (FLOW01).

3.3.3 Pseudonymization

We prompted Gemini 2.5 Flash to integrate a pseudonymization step using a second call. This involved a second call to the Azure OpenAI GPT-4.1 API with a tailored system prompt designed to identify and pseudonymize specified named entities while preserving the Markdown format. For details about the selected named entities, we refer to Section 3.1.1. The code used is available online via our GitHub Repository (see text footnote 1) (FLOW02).

3.3.4 EHR-synthesis co-developed with GenAI technology

Subsequently, we instructed Gemini 2.5 Flash to generate Python code that implements the synthetic data generation process. This process used the pseudonymized Markdown files as contextual examples for GPT-4.1, guided by the natural language prompts previously specified for the supervisor agent and worker agents (see Section 3.1.5). A comprehensive explanation of the code is provided, and it is publicly available as a downloadable Jupyter notebook from our GitHub Repository (see text footnote 1) (FLOW03).

3.3.5 Benchmark framework & analysis

Finally, Gemini 2.5 Flash was tasked with developing a programmatic approach to assess how closely synthetic data mirrors the informational and linguistic characteristics of pseudonymized real-world EHRs. This assessment strictly adheres to an evaluation framework comprising ten distinct metrics, as detailed in the next Section 3.4, to evaluate the indistinguishability of synthetic from real data. The code used is available online via our GitHub Repository (see text footnote 1) (FLOW04).

3.4 Quantitative assessment of synthetic data quality

To systematically assess the fidelity of GA-assisted SHDG, we recognized that no single metric would suffice. Therefore, we selected ten distinct metrics to jointly evaluate both the informational and linguistic qualities of clinical documents on both the individual document and corpus levels. These metrics, summarized in Tables 4, 5, were chosen to provide clear and quantifiable assessments that are informative for both AI specialists and clinical experts.

Table 4
www.frontiersin.org

Table 4. Metrics used to evaluate surface-level similarities.

Table 5
www.frontiersin.org

Table 5. Metrics used to evaluate linguistic and information level similarities.

Our framework goes beyond surface-level resemblance—such as basic structure—by also examining deeper linguistic and semantic properties. Specifically, we evaluate whether synthetic texts authentically mirror genuine EHRs in their phrasing, stylistic features, and conveyed meanings. To ensure this, we analyzed pooled corpora of pseudonymized real and synthetic data (Table 5), assessing similarities in word sequences, style, and intent.

Detailed implementation notes, equations and code for each metric are available in our public GitHub Repository (see text footnote 1) (FLOW04).

3.4.1 Document level assessment

Three different metrics —the original clinical documents, their pseudonymized counterparts, and synthetic documents generated through the GA-assisted SHDG process—were used to provide insight into the structural fidelity of the generated documents, ensuring that basic textual characteristics were faithfully reproduced in the synthetic samples. Assessment of averaged word count, average unique word count, and average document length (measured in characters) provided a straightforward means for comparing the overall size and composition of documents between the original, pseudonymized, and synthetic datasets.

Table 4 provides an overview of structural and linguistic metrics used to evaluate the fidelity of GA-assisted SHDG at the document level. It specifically focuses on comparing the original, pseudonymized, and synthetic documents in terms of word count, vocabulary diversity, and document length. The table offers guidance for interpreting whether the observed scores indicate that synthetic documents adequately replicate the structure and linguistic richness of the real data or reveal potential discrepancies.

Collectively, the document-level metrics of Table 4 allow for direct comparison of volume and informational content across the three datasets. Consistent differences in document length, word count, or vocabulary diversity between synthetic and real documents can signal problems in the data generation process, such as systematic under- or over-generation. Early detection of such discrepancies enables targeted improvements, ensuring that synthetic data more accurately reflects the completeness and verbosity found in authentic datasets.

3.4.2 Corpus level assessment

The inherent heterogeneity in EHRs—stemming from both the variety in patients and the diversity in documentation by healthcare professionals—can greatly impact comparability between real and synthetic data. Furthermore, as the aim of synthetic data is often to generate more data than is originally available (thus overcoming limited data availability), metrics other than pairwise comparison of individual documents are needed to evaluate the linguistic and informational similarities between datasets. To assess the actual comparability between the real pseudonymized and the synthetic clinical text, we pooled the individual documents into two corpora.

Table 5 provides a comprehensive set of metrics for assessing the fidelity of synthetic clinical text at the corpus level. Given the inherent variability in real-world electronic health records and the goal of creating synthetic datasets that closely resemble the originals, these metrics move beyond simple pairwise document comparisons to evaluate broader linguistic and informational features. The table outlines measures of textual diversity, vocabulary similarity, semantic alignment, and machine discernibility, each with its specific interpretation. Collectively, these metrics enable a holistic evaluation of how well the synthetic corpus replicates the complexity, nuance, and realism of the real clinical text, providing an in-depth view of corpus-level similarity across multiple dimensions.

3.5 Sample dataset description

The original dataset comprised N = 13 EHRs in PDF format, all relating to Dutch patients suffering from lower back pain. These real-world documents featured a combination of structured and unstructured text, including clinical notes, reports, and other pertinent patient information. To characterize these EHRs in terms of linguistic quality and informational content, a custom Python script was developed through GenAI-assisted co-development (see Section 3.3 for a detailed description).

For each file, various parameters were extracted and calculated, including storage size (in MB), textual content size (measured as total words, unique words after tokenization and lowercasing, and total characters as a measure of document length), and the primary language detected within the textual content. Structural elements such as the presence and count of tables, figures (images), and annotations were also identified. Additionally, Shannon entropy was computed at both the character and word level to quantify the average uncertainty or randomness in the text, thereby providing insight into its complexity and predictability. The canonical Jensen-Shannon Divergence (JSD) was calculated to compare the word distribution of each document with the overall word distribution across the dataset, reflecting the distinctiveness of each document’s language use. Finally, the average pointwise mutual information (PMI) for word bigrams was determined to assess the strength of association between commonly co-occurring words. Note, the use of PMI was inspired by the mutual information approach used in the neurophysiology study by van der Willigen et al. (2024), which measures the dependence between complex spectral-temporal sound representations. Both methods apply principles of information theory to quantify meaningful relationships related to NLP, though in distinct data domains (language vs. auditory processing) and at different scales (word pairs vs. neural coding of sound features).

4 Results

We start by reporting on a document-level assessment (Section 4.1), examining whether the generated data are contextually and semantically consistent with real EHRs. This is followed by a corpus-level assessment (Section 4.2) to assess whether our GA-assisted SHDG protocol adhered to established clinical data standards and preserved the statistical properties of the sample dataset (for details, see Section 3.5).

4.1 Document level assessment

The first 4 rows of Figure 4—Size (MB), Word Count, Unique Words, Document Length (Chars)—provide a structural element characterization of the original real-world, pseudonymized, and synthetic datasets, respectively. Each real-world EHR document—provided in PDF format—was analyzed for both linguistic quality and informational content. Structural document components—such as tables and images—were also quantified; across the original EHRs, between two and four tables were identified per file, while no figures (images) were detected in any document.

Figure 4
Violin plots comparing various metrics across three datasets: Real-World PDF, Pseudonymized Markdown, and Synthetic Markdown. Metrics shown are size in megabytes, word count, unique words, document length in characters, character entropy, word entropy, average pointwise mutual information, and Jensen-Shannon divergence distance. Each metric is color-coded and the plots illustrate distributions with density and outlier points.

Figure 4. Document level assessment using a holistic benchmark framework for quantitative evaluation of synthetic data quality (see Table 4). The figure presents a matrix of violin plots comparing the distributions of eight features—Size (MB), Word Count, Unique Words, Document Length (Chars), Character Entropy, Word Entropy, Average Pointwise Mutual Information (PMI), and Jensen-Shannon (JS) Distance—across three datasets: (A) real-world PDF (N = 13), (B) pseudonymized markdown (N = 13), and (C) synthetic markdown (N = 20). Note, each feature is encoded by a unique color for visual clarity (see Legend upper right side of the figure). Violin plots combine aspects of box plots and kernel density plots to provide a nuanced visualization of distributional characteristics. Specifically, the width of each violin at a given value represents the estimated probability density of the data at that value, as calculated by a kernel density estimator. This allows for the depiction of multimodality, skewness, and overall distributional shape, beyond summary statistics such as mean or quartiles. In this matrix, each subplot includes both the violin plot and overlaid scatter points indicating individual data instances, thereby facilitating both distributional and sample-level comparison of features across the three datasets.

When converting the original, real-world sample data from PDF to pseudonymised Markdown format, file sizes dropped noticeably. This is because PDF files retain a large amount of information—including embedded fonts, images, and detailed layout instructions—to ensure consistent appearance across devices. In contrast, Markdown files contain only the essential textual content and minimal formatting information, omitting media and complex layout data, which makes them much more compact (first row, Figure 4), However, during the pseudonymization phase, we added information to the documents (see Section 3.1.1). Consequently, the word count, and document length of the pseudonymized data increased somewhat (second row, Figure 4), reflecting the inclusion of additional descriptive tokens and labels in the file.

The overall structure of the synthetic data differed substantially from the pseudonymized data: on average, synthetic clinical notes were approximately 30% shorter in length (9,412 vs. 13,370). This reduction suggests that synthetic documents contained significantly less content per record, potentially omitting important clinical details or context. While real-world EHRs often include duplicate or redundant information (Nijor et al., 2022; Steinkamp et al., 2022), we observed that the synthetic data lacked such redundancy. This absence likely accounts for the notable decrease in word count and document length in the synthetic notes, even though the number of unique words remained similar between the two datasets.

At the per-document level, the mean bigram PMI was also slightly lower for synthetic documents (6.26) compared to pseudonymized documents (6.40); however, this difference was not statistically significant (Mann–Whitney U = 161.00, p = 0.2611). This suggests that although, on average, synthetic texts have somewhat weaker word pairings, individual documents do not consistently differ in bigram association strength, indicating that variability in bigram usage exists within both pseudonymized and synthetic documents.

4.2 Corpus level assessment

We compared our synthetic clinical text to the real-world, pseudonymized clinical notes at corpus level (Section 4.2) using informational and linguistic measures, including Shannon’s Entropy and average bigram PMI. The results are shown in Table 6.

Table 6
www.frontiersin.org

Table 6. Corpus level assessment: pseudonymized (Pseudo) versus synthetic (Synth) text corpora.

The corpus-level Shannon entropy quantifies the diversity and unpredictability of character and word usage within a corpus. For character-level entropy, the pseudonymized corpus (4.8565) exhibited a higher value than the synthetic corpus (4.6936), indicating that texts in the pseudonymized set utilize a greater variety of characters or employ characters in a less predictable manner. This suggests that the process of synthetically generating text may introduce constraints or redundancies at the character level, resulting in reduced diversity.

Conversely, word-level corpus entropy was higher in the synthetic corpus (9.9799) compared to the pseudonymized corpus (9.3543). This reflects a broader or less predictable word usage in the synthetic data, potentially attributable to the generative process introducing new combinations of words or emphasizing novelty. Thus, while synthetic data appears to be less varied at the character level, it is more varied at the word level than the original pseudonymized corpus.

Beyond overall corpus-level entropy, per-document analysis further clarifies differences in diversity and distribution between the pseudonymized and synthetic corpora. Mean per-document Shannon entropy at the character level was higher for the pseudonymized documents (4.8449) than for synthetic ones (4.6846), and this difference was statistically significant (Mann–Whitney U = 260.00, p < 0.001). This corroborates the corpus-level finding, indicating that, on an individual document basis, pseudonymized texts are consistently more diverse and less predictable regarding character usage than their synthetic counterparts.

Moreover, mean per-document Shannon entropy at the word level was significantly higher for synthetic documents (8.6939) than for pseudonymized ones (8.3075) (Mann–Whitney U = 0.00, p < 0.001). Thus, at the document level, synthetic texts exhibit greater unpredictability and a broader vocabulary than those found in the pseudonymized set, as was the case for corpus-level entropy.

For Average Bigram PMI, which quantifies the strength of association between word pairs by measuring how much more likely two words are to occur together than would be expected by chance, we observed that the synthetic corpus exhibited weaker word pair associations compared to the pseudonymized corpus. Specifically, the corpus-average bigram PMI was lower in the synthetic data (5.98) than in the pseudonymized data (6.92). This indicates that bigrams in synthetic texts are less strongly associated, or less conventional, than those present in the real-world clinical narratives. This finding likely reflects the inherent repetitiveness and predictability of authentic EHRs, in which common phraseology and co-occurring terms are documented repeatedly, thereby increasing the frequency and association strength of certain word pairs. In contrast, synthetically generated texts may introduce more varied or less natural co-occurrences, leading to overall weaker bigram associations.

In contrast to Shannon’s Entropy and average bigram PMI, metrics such as JSD, BLEU score, BERTScore, and classifier performance directly provide a comparative value as output (Table 6). While the synthetic data demonstrates several meaningful similarities to real data, there are also notable shortcomings.

The mean per-document Jensen-Shannon divergence (JSD) of word distributions relative to the combined corpus is higher in synthetic documents (0.2618) than in pseudonymized ones (0.2297), a statistically significant difference (Mann–Whitney U = 12.00, p < 0.001). This signifies that individual synthetic documents tend to diverge more from the aggregate corpus word distribution than pseudonymized documents, implying less conformity and potentially greater variation in how topics or vocabulary are expressed in synthetic data.

In surface-level text analysis, an n-gram refers to a sequence of n consecutive words—such as a bigram (two words) or a trigram (three words)—and comparing n-grams between documents helps reveal how closely their wording and phrasing align. Surface-level text similarity, as measured by the BLEU score (4.62 out of 100), was very low, indicating minimal n-gram overlap between synthetic and pseudonymized texts. This result bears out that the synthetic data are not merely replicating or closely paraphrasing real document phrases but are instead generating genuinely novel text content. While a low BLEU score might be viewed as a negative outcome in tasks requiring close mimicry, in the context of privacy-preserving data synthesis, it is encouraging. It demonstrates that the generative Agent workflow is not merely memorizing or reproducing existing expressions from the source corpus, but creating new, diverse language that reduces risks of information leakage.

Semantic similarity, as assessed by BERTScore, was moderate, with an F1 score of approximately 0.65 (precision: 0.64, recall: 0.66). Thus, although the synthetic data exhibits low surface-level overlap and increased lexical diversity compared to authentic EHR notes, it nonetheless preserves a substantial portion of the underlying clinical meaning and topical content. Such moderation in semantic overlap tells us that the sentences and phrases in the synthetic data are not directly copied or closely matched, word-for-word, with those in the original clinical records (authentic EHR notes). When we look at common sequences of words (n-grams), there is very little overlap between the two sets—meaning the synthetic data displays different combinations of words and sentences, rather than repeating those found in the real notes. In our use case of synthesizing clinical narratives, very high F1, precision, and recall measures would indicate exact copies of original data, whereas our aim was to increase textual diversity by adding real-world samples and clinical practice guidelines as knowledge bases for reasoning.

A machine learning classifier trained to distinguish between synthetic and pseudonymized (real) documents attained perfect discrimination, with both AUC and AUPRC scores of 1.00. This result shows that there are clear, easily learnable feature differences between the two datasets—particularly those reflected in TF-IDF representations. Among these, document length emerged as a primary distinguishing characteristic. Consequently, while the synthetic dataset demonstrates advantages such as content novelty and satisfactory conceptual coverage, its inability to realistically replicate document length results in synthetic documents being consistently and trivially separable from real ones. This highlights the need for improved modeling of document-level properties to enhance the realism and utility of synthetic clinical text. It also provides a clear path forward: simply adjusting the length of synthetic documents to match that of authentic clinical notes, or by deleting redundant information from the authentic clinical notes, should eliminate the primary feature the classifier uses to tell them apart. In doing so, the synthetic texts would likely become much harder for automated classifiers to distinguish from real ones, substantially improving their realism and the utility of the synthetic dataset.

Our corpus-level assessment—drawing on both Figure 4 and Table 6—shows that clinical text derived through GA-assisted SHDG are more predictable at the character level but surpass pseudonymized notes in word-level diversity and unpredictability. This pattern likely stems from artifacts or variability introduced during text generation, with important implications for the realism and utility of synthetic data. Overall, our findings highlight that while synthetic EHRs successfully avoids direct replication and privacy risks by producing novel word combinations, they also exhibit weaker conventional phrase associations and greater divergence from word distributions observed in the real corpus in.

5 Discussion

Our present work addresses a persistent barrier in digital health GenAI: the lack of accessible, interoperable, and privacy-preserving datasets that capture the diversity of real-world healthcare documentation (Eigenschink et al., 2023; Murtaza et al., 2023; Hernandez et al., 2025; Ibrahim et al., 2025; Loni et al., 2025). Building on our stepwise strategy for constructing a learning health system (LHS) (van Velzen et al., 2023), we argue that a fully integrated LHS will remain unattainable until these data challenges—especially in nursing and allied health (Tischendorf et al., 2025)—are resolved.

Echoing “On the Dangers of Stochastic Parrots” (Bender et al., 2021), we stress that healthcare must ask whether public cloud LLMs can be used in ways that are truly FAIR—Findable, Accessible, Interoperable, and Reusable—while mitigating risks such as high environmental and financial costs, opaque or biased data, and amplification of inequities. This requires curating and documenting high-quality datasets, aligning development with research and stakeholder values, and exploring approaches beyond ever-larger models.

A viable solution is the adoption of compact, fine-tunable SLMs. These small LLM alternatives can be deployed directly within healthcare facilities. Running locally not only reduces reliance on external cloud services but also lowers operational costs, decreases energy demands, and enhances data privacy. SLMs are increasingly capable of powering point-of-care applications—from clinical decision support to patient communication—while ensuring sensitive information remains within institutional boundaries (Schick and Schütze, 2021; Dibia et al., 2024; Garg et al., 2025; Kim et al., 2025; Xie et al., 2025).

By establishing a no-code, protocol for creating GA-assisted SHDG workflows—enabled by rapid prototyping platforms and further enhanced through fully automated, GenAI-assisted co-development—it is achievable to significantly lower the technical threshold for GenAI engagement. Leveraging the DSI stack as a generic blueprint architecture (Figure 1A), our methodological approach illustrates how modular, no-code frameworks can be systematically employed to streamline and democratize the creation of intelligent healthcare workflows.

In particular, the use of GA-assisted workflows to support clinical reasoning for nurse specialists—demonstrated through the “Nandalyse”11 tool at the 2025 ASCENDIO conference—represents a major shift in healthcare technology development and adoption in the Netherlands (Kumar, 2023; Jeong, 2025; Tischendorf et al., 2025). These novel GenAI tools enable clinicians and other non-technical users to easily create and customize workflows using simple, modular, drag-and-drop interfaces, significantly lowering the barriers to participation.

A key aspect of our protocol for GA-assisted SHDG workflows is the explicit bridging of the gap between end users—such as clinicians, quality officers, and researchers—and the technical developers responsible for constructing and maintaining AI systems (as detailed in Figure 1B). In healthcare, this chiasm is often perpetuated by differences in language, priorities, and familiarity with digital tools. The explicit integration of human-in-the-loop (Alemohammad et al., 2024) in no-code GA-assisted SHDG workflows not only supports iterative co-development (Li et al., 2023) but also ensures that the system remains transparent and grounded in real-world clinical needs and documentation standards (Chung et al., 2023).

Notably, our methodology fosters closer collaboration between users and developers by providing open-source GitHub repositories (see text footnote 4), which make design choices, workflow logic, and evaluation criteria more transparent and verifiable. However, realizing the full promise of these platforms will require continued investment in user education, ongoing refinement of documentation and support resources, and the cultivation of communities of practice around open-source synthetic data generation.

The successful adoption of privacy preserving GA-assisted SHDG workflows in clinical practice, healthcare professionals requires more than technical proficiency alone (Kumar, 2023; Chew and Ngiam, 2025; Jeong, 2025); they must also possess a comprehensive understanding of key data privacy principles—such as pseudonymization and de-identification—to ensure patient confidentiality is maintained (Drechsler and Haensch, 2024; Rujas et al., 2025). Especially, clinicians should be aware of the inherent limitations of LLMs, including potential risks of bias, hallucination, and model drift, all of which may impact the fidelity and safety of synthetic health data (Alemohammad et al., 2024; Loni et al., 2025). Harnad (2025), notes that LLMs rely on stochastic patterns by capturing statistical regularities rather than achieving genuine semantic understanding, meaning that even fluent and coherent output may be inaccurate, irrelevant or misleading. Shojaee et al. (2025) further caution that the apparent reasoning proficiency of such models can deteriorate markedly as problem complexity increases—an “illusion of thinking” with significant implications for clinical safety. Collectively, these observations reinforce the imperative for multidimensional evaluation frameworks that integrate both surface-level measures (e.g., document length, lexical overlap) and deep-level metrics (e.g., semantic alignment, diversity) to rigorously assess the fidelity and practical utility of synthetic narratives across diverse, high-stakes clinical contexts (Shannon, 1948; Post, 2018; Zhang et al., 2019).

Importantly, working with GenAI in Healthcare mandates a clear understanding of evaluation metrics (Eigenschink et al., 2023; Abdurahman et al., 2025; Ibrahim et al., 2025). Because no single metric can holistically capture the quality of synthetic clinical text, a comprehensive assessment should integrate document- and corpus-level metrics (for overview see Tables 4, 5, respectively): document length and average word count for surface features, entropy for textual diversity (Shannon, 1948), BLEU for lexical overlap (Post, 2018), and BERTScore for semantic alignment (Zhang et al., 2019). Together, these complementary metrics provide a nuanced perspective on both the linguistic and informative value of synthetic narratives. Importantly, high performance on one dimension (e.g., diversity as measured by entropy) does not necessarily translate to strong semantic faithfulness (as measured by BERTScore). This distinction is particularly relevant when evaluating synthetic data for diverse clinical contexts, such as physiotherapy documentation in high-risk or emotionally fraught scenarios (e.g., cardiac, oncology, or orthopedic surgery). Our results (Table 6) reinforce the necessity of using an ensemble of surface and deep metrics to robustly assess the utility and fidelity of synthetic clinical narratives, supporting their appropriate integration into research and practice.

Our work is not without limitations. We identified four main operational and technical challenges that must be addressed to advance GA-assisted SHDG workflows. First, comparative evaluation of on-premises versus cloud-based AI models (Table 1) is needed to optimize trade-offs between data privacy, computational performance, and cost (Schick and Schütze, 2021; Garg et al., 2025; Xie et al., 2025). Parameter tuning—including the adjustment of prompt temperature and chunk size—must be empirically refined to balance realism and diversity in generated outputs (Peeperkorn et al., 2024). Second, while our GA prompts were intentionally crafted to encourage adherence to established clinical practice guidelines to generate clinically relevant outputs, we acknowledge that real-world clinical practice often diverges from these standards. Notably, intentional non-adherence to guidelines has been observed in up to 65% of cases within EHRs, frequently attributable to specific patient factors such as contraindications, comorbidities, or individual preferences (Arts et al., 2016). Therefore, for GA-assisted SHDG workflows to remain practical and reflective of authentic clinical scenarios, addressing this variability is essential. Consequently, our prompt design may have inadvertently constrained the clinical diversity of the synthesized narratives—a limitation that may have been further amplified by the small sample size of our authentic clinical EHR documentation dataset (N = 13 PDF documents). Future SHDG workflows should explicitly address these limitations through testing various prompt engineering techniques to better capture the variability inherent in real-world clinical narratives, or alternatively, by increasing the size of or sample dataset to encompass a broader range of patient presentations, care contexts, and documentation styles and a wider range of intentional non-adherence to guidelines. Implementing these changes—refining prompt-engineering techniques and expanding the dataset—would improve the representativeness of the generated outputs and better align them with the complexities of actual clinical practice. Third, we identified additional prompt engineering issues (Figure 3). These included the inadvertent introduction of gender bias—all synthetic patients turned out to be female—and inconsistent handling of abbreviations—where sample and markdown files contained frequent abbreviations, but synthetic data did not. We also found another prompt engineering issue that resulted in a one-size-fits-all approach in the generated therapy plans and reduced therapy frequency variability. Document length discrepancies, often resulting from repetition in sample or markdown files, also require further exploration to ensure consistent structural fidelity across datasets. Recent research underscores that prompt engineering is not a value-neutral process and that different components of a prompt can vary significantly in their robustness and susceptibility to bias. Mei et al. (2025) provide a comprehensive survey of “context engineering” strategies for large language models, highlighting how the choice, structuring, and sequencing of contextual elements can systematically influence model outputs. Zheng et al. (2025) further demonstrate that individual prompt components—such as instructions, examples, or delimiters—exhibit heterogeneous adversarial robustness, meaning that some parts are more vulnerable to manipulation or unintended bias than others. Related work on priming effects shows that the initial context or examples provided to a model can strongly condition its subsequent responses, amplifying or dampening biases and shaping output diversity (Zhao et al., 2021; Gallegos et al., 2024). Huang et al. (2025) demonstrate that targeted priming can exploit intrinsic weaknesses in large language models, revealing latent vulnerabilities that may not be apparent under standard prompting conditions. Meinke et al. (2024) further reveal that frontier-scale models are capable of sophisticated in-context behaviors, sometimes strategically adapting to earlier cues in ways that can subtly steer reasoning and decision-making. In addition, Gallegos et al. (2024) synthesize evidence that such biases can emerge not only from pre-training data but also from contextual framing and priming during inference. Together, these findings suggest that the gender bias, abbreviation inconsistencies, and homogenized therapy plans observed in our study may stem not only from prompt content but also from the structural composition, priming effects, and resilience of the prompts themselves. This reinforces the need for systematic evaluation and refinement of both prompt components and priming strategies to mitigate bias and improve variability in generated outputs. Fourth, the normalization of clinical text presents a significant methodological challenge, as variations in terminology—for instance, describing the same condition as “sciatica” versus “lumbosacral radicular pain syndrome”—often reflect individual practitioner preferences as well as institutional conventions (Suominen et al., 2013; U.S.-National-Library-of-Medicine, 2024). This semantic variability can be quantitatively monitored using mutual information-based evaluation metrics (see Section 3.4); in our study, we applied the average bigram pointwise mutual information (PMI) metric. However, the robustness of this metric across more diverse datasets and practitioners from different clinical specialties remains to be fully validated. Thus, relying solely on either surface-level metrics or deep semantic measures is insufficient. Instead, a comprehensive evaluation of the faithfulness of SHDG demands the integration of both approaches (Eigenschink et al., 2023; Smolyak et al., 2024; Chen et al., 2025b; Ibrahim et al., 2025; Scherr et al., 2025).

6 Conclusion

Based on our findings, we recommend prioritizing the continued development and refinement of GA-assisted SHDG-workflows, ensuring that these toolchains remain accessible, transparent, and customizable for a diverse range of clinical users and researchers through the provision of open-source GitHub repositories (see text footnote 4). Our recommendations align with ongoing efforts by the RUAS Healthcare DataLab to advance innovation in care through collaborative, technology-driven solutions that integrate open, adaptable tools into diverse healthcare contexts.12 In parallel, our talent program fosters and equips RUAS students with the skills and expertise required to become proficient data science professionals, thereby strengthening the capacity for data-driven innovation within the Dutch healthcare sector. To maximize impact, future initiatives should emphasize robust user education on essential data privacy concepts—such as pseudonymization and de-identification—and foster a deeper understanding of both the capabilities and limitations of no-code GenAI technologies, particularly regarding risks like bias, hallucination, and model drift. Comprehensive evaluation frameworks should integrate both surface-level and deep semantic metrics to robustly assess the linguistic and clinical fidelity of synthetic narratives, while methodological improvements—including enhanced prompt engineering to address issues of bias, guideline adherence, and structural consistency, as well as adherence to normalization standards—will be essential in capturing the diversity inherent in real-world clinical documentation. Comparative assessments of on-premises versus cloud-based deployments, plus empirical tuning of generative parameters, are also advised to optimize privacy, cyber security, cost, and performance trade-offs for varied healthcare settings. Finally, building active communities of practice and cultivating collaborative, iterative engagement between end-users and developers will be critical to realizing truly FAIR (Mons et al., 2017), AI-ready LHS that can adapt to the complexities and dynamic requirements of modern clinical environments (Huerta et al., 2023; Verhulst et al., 2025).

Our protocolization of privacy-preserving, GA-assisted SHDG workflows underscores the critical importance of maintaining transparency by keeping humans actively involved in the process—a principle known as human-in-the-loop. Specifically, we standardized the use of a modular, no-code system in which every step—from designing AI prompts, to providing data inputs, to defining how outputs are applied—can be easily inspected, understood, and refined without programming expertise. This approach enables users to trace how the LLM generates clinical narratives and to make step-by-step improvements over time. In parallel, the integration of an open-source GitHub repository (see text footnote 4), anchored by our modular DSI Stack (Figure 1A), supports the design and deployment of secure, reproducible, and scalable GA-assisted SHDG workflows across both on-premises and public cloud environments. By incorporating Docker-based containerization and API key–secured inference endpoints via Hugging Face Spaces, we ensure controlled access, consistent performance, and community-driven enhancement, while shielding users from underlying infrastructure complexity and maintaining full transparency for healthcare and research applications.

Data availability statement

The sample datasets referenced in this article are not publicly available, as they are proprietary to MediFit Bewegingscentrum Oss and the University of Applied Sciences Rotterdam. At present, approval to use these datasets beyond the university has not been granted. Individuals interested in accessing the sample data are kindly asked to contact Mark van Velzen at bS52YW4udmVsemVuQGhyLm5s with their request. The original contributions presented in the study—source code, synthetic datasets and prompts—are included in the article/supplementary material. Toolchain implementation details are provided to facilitate reproducibility and to encourage non-AI experts, such as healthcare workers, to confidently explore, adapt, and apply these workflows within their own professional contexts. Further inquiries can be directed to the corresponding author.

Ethics statement

Ethical approval was not required for this study involving humans, in accordance with local legislation and institutional requirements. The data were initially collected for the purpose of providing care. Prior to data extraction, all participants provided written informed consent, in accordance with the Declaration of Helsinki (World Medical Association, 2025). Following informed consent, the data were anonymized for research purposes.

Author contributions

MV: Writing – original draft, Formal analysis, Project administration, Methodology, Data curation, Writing – review & editing, Conceptualization, Investigation, Resources, Validation. RW: Writing – original draft, Formal analysis, Resources, Methodology, Software, Project administration, Visualization, Data curation, Writing – review & editing, Validation, Investigation, Conceptualization, Supervision. VB: Investigation, Validation, Writing – review & editing, Conceptualization, Methodology, Formal analysis, Writing – original draft. HG-W: Methodology, Conceptualization, Validation, Writing – review & editing, Writing – original draft, Investigation. EJ: Writing – review & editing, Conceptualization. SL: Resources, Writing – review & editing. MiW: Software, Writing – review & editing, Methodology, Validation, Conceptualization, Formal analysis. MaW: Writing – review & editing, Validation, Software, Methodology. GR: Software, Writing – review & editing. RM: Software, Writing – review & editing. SS: Software, Writing – review & editing. LH: Conceptualization, Writing – review & editing, Software, Methodology. TH: Conceptualization, Writing – review & editing. NM: Writing – review & editing, Conceptualization. MS: Conceptualization, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Acknowledgments

Our sincere thanks go to Maurice A.C. van den Dobbelsteen—innovation manager AI & eHealth at KCZI and HR-DataLab Healthcare coordinator—for his pivotal role in establishing our DataLab at RUAS. This project was truly a team effort, and his inspiring guidance, collaborative spirit, and insightful discussions were essential in enabling and shaping our research and Data Science talent program.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that Gen AI was used in the creation of this manuscript. Generative tools, specifically Gemini Advanced, Perplexity, and Deep Research Pro, were utilized in developing the manuscript drafts to enhance the quality of human-written text produced by non-native English-speaking authors. The use of these tools was fully in accordance with relevant legal requirements and ethical guidelines, including Article 12 GDPR (transparency regarding data processing), Article 5 GDPR (purpose limitation and data minimization), Article 13 AI ACT (informing users about interactions with AI systems), and Article 50 AI ACT (transparency requirements for AI providers and users). Furthermore, all AI-generated content was carefully reviewed and edited by human authors to ensure accuracy, relevance, and adherence to academic and ethical standards. Importantly, this process was conducted in compliance with the principles set forth in the “Nederlandse gedragscode wetenschappelijke integriteit” (Knaw et al., 2018).

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

References

Abdurahman, S., Salkhordeh Ziabari, A., Moore, A. K., Bartels, D. M., and Dehghani, M. (2025). A primer for evaluating large language models in social-science research. Adv. Methods Pract. Psychol. Sci. 8:25152459251325174. doi: 10.1177/25152459251325174

Crossref Full Text | Google Scholar

Abhishek, M. K., and Rao, D. R. (2021). “Framework to secure docker containers” in Fifth world conference on smart trends in systems security and sustainability (WorldS4) (London, UK: IEEE), 152–156. doi: 10.1109/WorldS451998.2021.9514041

Crossref Full Text | Google Scholar

Ait, A., Cánovas Izquierdo, J. L., and Cabot, J. (2025). On the suitability of hugging face hub for empirical studies. Empir. Softw. Eng. 30, 1–48. doi: 10.1007/s10664-024-10608-8

Crossref Full Text | Google Scholar

Alemohammad, S., Casco-Rodriguez, J., Luzi, L., Humayun, A. I., Babaei, H., LeJeune, D., et al. (2024). Self-consuming generative models go mad, in: International conference on learning representations (ICLR), (Vienna, AT). doi: 10.48550/arXiv.2307.01850

Crossref Full Text | Google Scholar

Alsentzer, E., Murphy, J. R., Boag, W., Weng, W.-H., Jin, D., Naumann, T., et al. (2019). “Publicly available clinical BERT embeddings” in Proceedings of the 2nd clinical natural language processing workshop (Minneapolis, MN: Association for Computational Linguistics), 72–78. doi: 10.18653/v1/W19-1909

Crossref Full Text | Google Scholar

Arts, D. L., Voncken, A. G., Medlock, S., Abu-Hanna, A., and van Weert, H. C. (2016). Reasons for intentional guideline non-adherence: a systematic review. Int. J. Med. Inform. 89, 55–62. doi: 10.1016/j.ijmedinf.2016.02.009

PubMed Abstract | Crossref Full Text | Google Scholar

Baowaly, M. K., Lin, C. C., Liu, C. L., and Chen, K. T. (2019). Synthesizing electronic health records using improved generative adversarial networks. J. Am. Med. Inform. Assoc. 26, 228–241. doi: 10.1093/jamia/ocy142

PubMed Abstract | Crossref Full Text | Google Scholar

Barrault, L., Duquenne, P.-A., Elbayad, M., Kozhevnikov, A., Alastruey, B., Andrews, P., et al. (2024). Large concept models: language modeling in a sentence representation space. arXiv [Preprint]. doi: 10.48550/arXiv.2412.08821

Crossref Full Text | Google Scholar

Beam, A.L., Kompa, B., Schmaltz, A., Fried, I., Weber, G., Palmer, N., et al. (2020). "Clinical concept embeddings learned from massive sources of multimodal medical data", in: Pacific symposium on Biocomputing 2020, (Singapore: World Scientific), 25:295–306

Google Scholar

Bender, E. M., Gebru, T., McMillan-Major, A., and Shmitchell, S. (2021). "On the dangers of stochastic parrots: can language models be too big?", in: Proceedings of the 2021 ACM conference on fairness, accountability, and transparency: Association for Computing Machinery, 610–623. doi: 10.1145/3442188.3445922

Crossref Full Text | Google Scholar

Bommasani, R., Hudson, D. A., Adeli, E., Altman, R., Arora, S., von Arx, S., et al. (2021). On the opportunities and risks of foundation models. arXiv [Preprint]. doi: 10.48550/arXiv.2108.07258

Crossref Full Text | Google Scholar

Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., et al. (2020). Language models are few-shot learners, in: Proceedings of the 34th international conference on neural information processing systems, (NY, USA: Curran Associates Inc.). doi: 10.5555/3495724.3495883

Crossref Full Text | Google Scholar

Busch, F., Kather, J. N., Johner, C., Moser, M., Truhn, D., Adams, L. C., et al. (2024). Navigating the European union artificial intelligence act for healthcare. NPJ Digit Med 7:210. doi: 10.1038/s41746-024-01213-6

PubMed Abstract | Crossref Full Text | Google Scholar

Cannon, J., and Lucci, S. (2010). Transcription and EHRS. Benefits of a blended approach. J. AHIMA 81, 36–40.

Google Scholar

Casper, S., Bailey, L., Hunter, R., Ezell, C., Cabalé, E., Gerovitch, M., et al. (2025). The AI agent index. arXiv [Preprint]. doi: 10.48550/arXiv.2502.01635

Crossref Full Text | Google Scholar

Chan, A., Salganik, R., Markelius, A., Pang, C., Rajkumar, N., Krasheninnikov, D., et al. (2023). “Harms from increasingly agentic algorithmic systems” in 2023 ACM conference on fairness accountability and transparency (New York, NY: Association for Computing Machinery), 651–666. doi: 10.1145/3593013.3594033

Crossref Full Text | Google Scholar

Chen, Q., Hu, Y., Peng, X., Xie, Q., Jin, Q., Gilson, A., et al. (2025b). Benchmarking large language models for biomedical natural language processing applications and recommendations. Nat. Commun. 16:3280. doi: 10.1038/s41467-025-56989-2

PubMed Abstract | Crossref Full Text | Google Scholar

Chen, B., Zhang, Z., Langrené, N., and Zhu, S. (2025a). Unleashing the potential of prompt engineering for large language models. Patterns 6:101260. doi: 10.1016/j.patter.2025.101260

PubMed Abstract | Crossref Full Text | Google Scholar

Chew, B.-H., and Ngiam, K. Y. (2025). Artificial intelligence tool development: what clinicians need to know? BMC Med. 23:244. doi: 10.1186/s12916-025-04076-0

PubMed Abstract | Crossref Full Text | Google Scholar

Chojnacki, B. (2025). The ultimate guide to selecting the right large language model for [Online]. dsstream. Available online at: https://www.dsstream.com/post/the-ultimate-guide-to-selecting-the-right-large-language-model-for (Accessed May 11, 2025)

Google Scholar

Chung, J., Kamar, E., and Amershi, S. (2023). “Increasing diversity while maintaining accuracy: text data generation with large language models and human interventions” in Proceedings of the 61st annual meeting of the ACL (volume 1: Long papers) (Cambridge, MA: Association for Computational Linguistics (ACL)), 575–593. doi: 10.18653/v1/2023.acl-long.34

Crossref Full Text | Google Scholar

Coveney, P. V., and Succi, S. (2025). The wall confronting large language models. arXiv [Preprint]. doi: 10.48550/arXiv.2507.19703

Crossref Full Text | Google Scholar

Das, A. B., Ahmed, S., and Sakib, S. K. (2025). Hallucinations and key information extraction in medical texts: a comprehensive assessment of open-source large language models. arXiv [Preprint]. doi: 10.48550/arXiv.2504.19061

Crossref Full Text | Google Scholar

Daull, X., Bellot, P., Bruno, E., Martin, V., and Murisasco, E. (2023). Complex qa and language models hybrid architectures, survey. arXiv [Preprint]. doi: 10.48550/arXiv.2302.09051

Crossref Full Text | Google Scholar

DeepMind (2025). Alphaevolve: A gemini-powered coding agent for designing advanced algorithms [Online]. Available online at: https://deepmind.google/discover/blog/alphaevolve-a-gemini-powered-coding-agent-for-designing-advanced-algorithms/ (Accessed May 29, 2025).

Google Scholar

Dibia, V., Chen, J., Bansal, G., Syed, S., Fourney, A., Zhu, E., et al. (2024). Autogen studio: a no-code developer tool for building and debugging multi-agent systems, in: Proceedings of the 2024 conference on empirical methods in natural language processing: System demonstrations, (Stroudsburg, PA: Association for Computational Linguistics), 72–79. doi: 10.18653/v1/2024.emnlp-demo.8

Crossref Full Text | Google Scholar

Doan, S., Conway, M., Phuong, T. M., and Ohno-Machado, L. (2014). “Natural language processing in biomedicine; a unified system architecture overview” in Clinical bioinformatics. Methods in molecular biology. ed. R. Trent (New York, NY: Humana Press), 275–294.

Google Scholar

Drechsler, J., and Haensch, A.-C. (2024). 30 years of synthetic data. Stat. Sci. 39, 221–242. doi: 10.1214/24-STS927

Crossref Full Text | Google Scholar

Driehuis, F., Woudenberg-Hulleman, I., Verhof-van Westing, I. M., Geurkink, H., Hartstra, L., Trouw, M., et al. (2019). Verantwoording en toelichting kngf-richtlijn fysiotherapeutische dossiervoering 2019 [Online]. Amersfoort: Koninklijk Nederlands Genootschap voor Fysiotherapie (KNGF). Available online at: https://www.kngf.nl/app/uploads/2024/09/fysiotherapeutische-dossiervoering-2019-verantwoording-en-toelichting-versie-1.1.pdf (Accessed May 11, 2025).

Google Scholar

Eigenschink, P., Reutterer, T., Vamosi, S., Vamosi, R., Sun, C., and Kalcher, K. (2023). Deep generative models for synthetic data: a survey. IEEE Access 11, 47304–47320. doi: 10.1109/ACCESS.2023.3275134

Crossref Full Text | Google Scholar

European Parliament and Council. (2024). Regulation (EU) 2024/1689 of the European parliament and of the council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending certain union legislative acts (artificial intelligence act) [online]. Brussels: Official Journal of the European Union, L, 202. Available online at: https://eur-lex.europa.eu/eli/reg/2024/1689/oj/eng (Accessed May 15, 2025).

Google Scholar

Fu, Y., Mai, L., and Ustiugov, D. (2025). AI goes serverless: are systems ready? [online]. Sigops. Available online at: https://www.sigops.org/2025/ai-goes-serverless-are-systems-ready/ (Accessed May 11, 2025).

Google Scholar

Gallegos, I. O., Rossi, R. A., Barrow, J., Tanjim, M. M., Kim, S., Dernoncourt, F., et al. (2024). Bias and fairness in large language models: a survey. Comput. Linguist. 50, 1097–1179. doi: 10.1162/coli_a_00524

Crossref Full Text | Google Scholar

Garg, M., Raza, S., Rayana, S., Liu, X., and Sohn, S. (2025). The rise of small language models in healthcare: a comprehensive survey. arXiv [Preprint]. doi: 10.48550/arXiv.2504.17119

Crossref Full Text | Google Scholar

Goodfellow, I., Bengio, Y., and Courville, A. (2018). Deep learning. Cambridge, MA: MIT press.

Google Scholar

Goyal, M., and Mahmoud, Q. H. (2024). A systematic review of synthetic data generation techniques using generative AI. Electronics 13:3509. Available online at: https://www.mdpi.com/2079-9292/13/17/3509

Google Scholar

Gridach, M., Nanavati, J., Abidine, K. Z. E., Mendes, L., and Mack, C. (2025). Agentic ai for scientific discovery: a survey of progress, challenges, and future directions. arXiv [Preprint]. doi: 10.48550/arXiv.2503.08979

Crossref Full Text | Google Scholar

Günther, M., Mohr, I., Williams, D. J., Wang, B., and Xiao, H. (2024). Late chunking: contextual chunk embeddings using long-context embedding models. arXiv [Preprint]. doi: 10.48550/arXiv.2409.04701

Crossref Full Text | Google Scholar

Gupta, S. (2025). The rise of serverless AI: transforming machine learning deployment. Eur. J. Comput. Sci. Inf. Technol. 13, 45–67. doi: 10.37745/ejcsit.2013/vol13n54567

Crossref Full Text | Google Scholar

Han, T., Adams, L. C., Papaioannou, J.-M., Grundmann, P., Oberhauser, T., Löser, A., et al. (2023). Medalpaca – an open-source collection of medical conversational AI models and training data. arXiv [Preprint]. doi: 10.48550/arXiv.2304.08247

Crossref Full Text | Google Scholar

Harnad, S. (2025). Language writ large: LLMs, chatgpt, meaning, and understanding. Front. Artif. Intell. 7:1490698. doi: 10.3389/frai.2024.1490698

PubMed Abstract | Crossref Full Text | Google Scholar

Hart, E. M., Barmby, P., LeBauer, D., Michonneau, F., Mount, S., Mulrooney, P., et al. (2016). Ten simple rules for digital data storage. PLoS Comput. Biol. 12:e1005097. doi: 10.1371/journal.pcbi.1005097

PubMed Abstract | Crossref Full Text | Google Scholar

Haug, C. J. (2018). Turning the tables – the new european general data protection regulation. N. Engl. J. Med. 379, 207–209. doi: 10.1056/NEJMp1806637

PubMed Abstract | Crossref Full Text | Google Scholar

Hechler, E., Weihrauch, M., and Wu, Y. (2023). “Evolution of data architecture” in Data fabric and data mesh approaches with AI (Berkeley, CA: Apress), 3–15. doi: 10.1007/978-1-4842-9253-2_1

Crossref Full Text | Google Scholar

Hernandez, M., Osorio-Marulanda, P. A., Catalina, M., Loinaz, L., Epelde, G., and Aginako, N. (2025). Comprehensive evaluation framework for synthetic tabular data in health: Fidelity, utility and privacy analysis of generative models with and without privacy guarantees. Front. Digit. Health 7:1576290. doi: 10.3389/fdgth.2025.1576290

PubMed Abstract | Crossref Full Text | Google Scholar

Hettiarachchi, I. (2025). Exploring generative AI agents: architecture, applications, and challenges. J. Artif. Intell. Gen. Sci. 8, 105–127. doi: 10.60087/jaigs.v8i1.350

Crossref Full Text | Google Scholar

Hoofnagle, C. J., Van Der Sloot, B., and Borgesius, F. Z. (2019). The European Union general data protection regulation: what it is and what it means. Inf. Commun. Technol. Law 28, 65–98. doi: 10.1080/13600834.2019.1573501

Crossref Full Text | Google Scholar

Huang, Y., Zhan, R., Wong, D. F., Chao, L. S., and Tao, A. (2025). Intrinsic model weaknesses: how priming attacks unveil vulnerabilities in large language models. Assoc. Comput. Ling., 1405–1425. doi: 10.18653/v1/2025.findings-naacl.77

Crossref Full Text | Google Scholar

Huerta, E. A., Blaiszik, B., Brinson, L. C., Bouchard, K. E., Diaz, D., Doglioni, C., et al. (2023). Fair for AI: an interdisciplinary and international community building perspective. Sci Data 10:487. doi: 10.1038/s41597-023-02298-6

PubMed Abstract | Crossref Full Text | Google Scholar

Ibrahim, M., Khalil, Y. A., Amirrajab, S., Sun, C., Breeuwer, M., Pluim, J., et al. (2025). Generative AI for synthetic data across multiple medical modalities: a systematic review of recent developments and challenges. Comput. Biol. Med. 189:109834. doi: 10.1016/j.compbiomed.2025.109834

PubMed Abstract | Crossref Full Text | Google Scholar

Inoue, K. (2024). A practitioner’s guide to selecting large language models for your business needs [Online]. Veritone. Available online at: https://www.veritone.com/blog/a-practitioners-guide-to-selecting-large-language-models-for-your-business-needs/ (Accessed May 11, 2025].

Google Scholar

Jeong, C. (2025). Beyond text: implementing multimodal large language model-powered multi-agent systems using a no-code platform. J. Intell. Inf. Syst. 31, 191–231. doi: 10.13088/jiis.2025.31.1.191

Crossref Full Text | Google Scholar

Jin, M., Sang-Min, C., and Gun-Woo, K. (2025). Comcare: a collaborative ensemble framework for context-aware medical named entity recognition and relation extraction. Electronics 14:328. doi: 10.3390/electronics14020328

Crossref Full Text | Google Scholar

Kaplan, J., McCandlish, S., Henighan, T., Brown, T. B., Chess, B., Child, R., et al. (2020). Scaling laws for neural language models. arXiv [Preprint]. doi: 10.48550/arXiv.2001.08361

Crossref Full Text | Google Scholar

Karpathy, A. (2025). Vibe coding [Online]. Available online at: https://x.com/karpathy/status/1886192184808149383 (Accessed May 12, 2025).

Google Scholar

Kim, H., Hwang, H., Lee, J., Park, S., Kim, D., Lee, T., et al. (2025). Small language models learn enhanced reasoning skills from medical textbooks. npj Digital Medicine 8:240. doi: 10.1038/s41746-025-01653-8

PubMed Abstract | Crossref Full Text | Google Scholar

KnawNFU, NWO, TO Federatie, Vereniging HogescholenVsnu (2018). Nederlandse gedragscode wetenschappelijke integriteit. V1. doi: 10.17026/dans-2cj-nvwu

Crossref Full Text | Google Scholar

Kochanowska, M., and Gagliardi, W. R.with reference to Jonathan, B (2022). “The double diamond model: in pursuit of simplicity and flexibility” in Perspectives on design ii. eds. D. Raposo, J. Neves, and J. Silva (Cham: Springer International Publishing), 19–32.

Google Scholar

Krishnakumar, A., Ogras, U., Marculescu, R., Kishinevsky, M., and Mudge, T. (2023). Domain-specific architectures: research problems and promising approaches. ACM Trans. Embed. Comput. Syst. 22, 1–26. doi: 10.1145/3563946

Crossref Full Text | Google Scholar

Kumar, A. (2023). 5 best Flowise alternatives in 2024 to build AI agents [online]. Available online at: https://blog.fabrichq.ai/6-best-flowise-alternatives-in-2024-to-build-ai-agents-8af9cb572449 (Accessed March 24, 2025).

Google Scholar

LeCun, Y., Bengio, Y., and Hinton, G. (2015). Deep learning. Nature 521, 436–444. doi: 10.1038/nature14539

PubMed Abstract | Crossref Full Text | Google Scholar

Lee, J., Yoon, W., Kim, S., Kim, D., Kim, S., So, C. H., et al. (2020). Biobert: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 36, 1234–1240. doi: 10.1093/bioinformatics/btz682

PubMed Abstract | Crossref Full Text | Google Scholar

Li, Z., Zhu, H., Lu, Z., and Yin, M. (2023). “Synthetic data generation with large language models for text classification: potential and limitations” in Proceedings of the 2023 conference on empirical methods in natural language processing. eds. H. Bouamor, J. Pino, and K. Bali (Cambridge, MA: Association for Computational Linguistics (ACL)), 10443–10461. doi: 10.18653/v1/2023.emnlp-main.647

Crossref Full Text | Google Scholar

Little, R. J. (1993) Statistical analysis of masked data. J. Off. Stat., 9: 407–426. Available online at: https://www.imstat.org/publications/sts/sts_39_2/sts_39_2.pdf

Google Scholar

Liu, Y., Acharya, U. R., and Tan, J. H. (2025). Preserving privacy in healthcare: a systematic review of deep learning approaches for synthetic data generation. Comput. Methods Prog. Biomed. 260:108571. doi: 10.1016/j.cmpb.2024.108571

PubMed Abstract | Crossref Full Text | Google Scholar

Loni, M., Poursalim, F., Asadi, M., and Gharehbaghi, A. (2025). A review on generative AI models for synthetic medical text, time series, and longitudinal data. npj Digit. Med. 8:281. doi: 10.1038/s41746-024-01409-w

Crossref Full Text | Google Scholar

Mayo, M. (2025). Feel the vibe: Why AI-dependent coding isn’t the enemy (or is it?) [Online]. Available online at https://www.kdnuggets.com/feel-the-vibe-why-ai-dependent-coding-isnt-the-enemy-or-is-it (Accessed May 15, 2025).

Google Scholar

Mei, L., Yao, J., Ge, Y., Wang, Y., Bi, B., Cai, Y., et al. (2025). A survey of context engineering for large language models. arXiv [Preprint]. doi: 10.48550/arXiv.2507.13334

Crossref Full Text | Google Scholar

Meinke, A., Schoen, B., Scheurer, J., Balesni, M., Shah, R., and Hobbhahn, M. (2024). Frontier models are capable of in-context scheming. arXiv [Preprint]. doi: 10.48550/arXiv.2412.04984

Crossref Full Text | Google Scholar

Meng, X.-L. (2021). Building data science infrastructures and infrastructural data science. Harv. Data Sci. Rev. 3. doi: 10.1162/99608f92.abfa0e70

Crossref Full Text | Google Scholar

Meystre, S. M., Lovis, C., Burkle, T., Tognola, G., Budrionis, A., and Lehmann, C. U. (2017). Clinical data reuse or secondary use: current status and potential future progress. Yearb. Med. Inform. 26, 38–52. doi: 10.15265/IY-2017-007

PubMed Abstract | Crossref Full Text | Google Scholar

Mons, B., Neylon, C., Velterop, J., Dumontier, M., da Silva Santos, L. O. B., and Wilkinson, M. D. (2017). Cloudy, increasingly fair; revisiting the fair data guiding principles for the European open science cloud. Inf. Serv. Use 37, 49–56. doi: 10.3233/isu-170824

Crossref Full Text | Google Scholar

Morris, J. X., Sitawarin, C., Guo, C., Kokhlikyan, N., Suh, G. E., Rush, A. M., et al. (2025). How much do language models memorize? arXiv [Preprint]. doi: 10.48550/arXiv.2505.24832

Crossref Full Text | Google Scholar

Murtaza, H., Ahmed, M., Khan, N. F., Murtaza, G., Zafar, S., and Bano, A. (2023). Synthetic data generation: state of the art in health care domain. Comput Sci Rev 48:100546. doi: 10.1016/j.cosrev.2023.100546

Crossref Full Text | Google Scholar

Negro-Calduch, E., Azzopardi-Muscat, N., Krishnamurthy, R. S., and Novillo-Ortiz, D. (2021). Technological progress in electronic health record system optimization: systematic review of systematic literature reviews. Int. J. Med. Inform. 152:104507. doi: 10.1016/j.ijmedinf.2021.104507

PubMed Abstract | Crossref Full Text | Google Scholar

Nijor, S., Rallis, G., Lad, N., and Gokcen, E. (2022). Patient safety issues from information overload in electronic medical records. J. Patient Saf. 18, e999–e1003. doi: 10.1097/PTS.0000000000001002

PubMed Abstract | Crossref Full Text | Google Scholar

Novikov, A., Vu, N., Eisenberger, M., Dupont, E., Huang, P., Wagner, A.Z., et al. (2025). Alphaevolve: A coding agent for scientific and algorithmic discovery. [online] [Preprint]. Available online at: https://storage.googleapis.com/deepmind-media/DeepMind.com/Blog/alphaevolve-a-gemini-powered-coding-agent-for-designing-advanced-algorithms/AlphaEvolve.pdf (Accessed May 09, 2025).

Google Scholar

Ohse, J., Hadžić, B., Mohammed, P., Peperkorn, N., Danner, M., Yorita, A., et al. (2024). Zero-shot strike: testing the generalisation capabilities of out-of-the-box llm models for depression detection. Comput. Speech Lang. 88:101663. doi: 10.1016/j.csl.2024.101663

Crossref Full Text | Google Scholar

OpenAI. (2025). Introducing gpt-4.1 in the api [Online]. Available online at: https://openai.com/index/gpt-4-1/ (Accessed May 13, 2025).

Google Scholar

Orru, G., Piarulli, A., Conversano, C., and Gemignani, A. (2023). Human-like problem-solving abilities in large language models using chatgpt. Front Artif Intell 6:1199350. doi: 10.3389/frai.2023.1199350

PubMed Abstract | Crossref Full Text | Google Scholar

Park, J. S., O'Brien, J., Cai, C. J., Morris, M. R., Liang, P., and Bernstein, M. S. (2023). Generative agents: interactive simulacra of human behavior, in: Proceedings of the 36th annual ACM symposium on user Interface software and technology (UIST '23), San Francisco, CA: ACM, 1–22. doi: 10.1145/3586183.3606763

Crossref Full Text | Google Scholar

Peeperkorn, M., Kouwenhoven, T., Brown, D., and Jordanous, A. (2024). Is temperature the creativity parameter of large language models?, in: Proceedings of the 15th international conference on computational creativity (ICCC’24), (Coimbra, Portugal: Association for Computational Creativity). doi: 10.48550/arXiv.2405.00492

Crossref Full Text | Google Scholar

Pezoulas, V. C., Zaridis, D. I., Mylona, E., Androutsos, C., Apostolidis, K., Tachos, N. S., et al. (2024). Synthetic data generation methods in healthcare: a review on open-source tools and methods. Comput. Struct. Biotechnol. J. 23, 2892–2910. doi: 10.1016/j.csbj.2024.07.005

PubMed Abstract | Crossref Full Text | Google Scholar

Piccialli, F., Chiaro, D., Sarwar, S., Cerciello, D., Qi, P., and Mele, V. (2025). Agentai: a comprehensive survey on autonomous agents in distributed AI for industry 4.0. Expert Syst. Appl. 291:128404. doi: 10.1016/j.eswa.2025.128404

Crossref Full Text | Google Scholar

Post, M. (2018). A call for clarity in reporting bleu scores, in: Proceedings of the 3rd conference on machine translation: Research papers, (Stroudsburg, PA: Association for Computational Linguistics), 186–191. doi: 10.18653/v1/W18-6319

Crossref Full Text | Google Scholar

Priebe, T., Neumaier, S., and Markus, S. (2021). "Finding your way through the jungle of big data architectures", in: 2021 IEEE international conference on big data (big data), (New York, NY: IEEE), 5994–5996. doi: 10.1109/BigData52589.2021.9671862

Crossref Full Text | Google Scholar

Qiu, X., Wang, H., Tan, X., Qu, C., Xiong, Y., Cheng, Y., et al. (2024). Towards collaborative intelligence: propagating intentions and reasoning for multi-agent coordination with large language models. arXiv [Preprint]. doi: 10.48550/arXiv.2407.12532

Crossref Full Text | Google Scholar

QuantSpark. (2023). Choosing the right llm: A guide for decision-makers. Available online at: https://quantspark.ai/blogs/2023/10/31/ai-comparison-guide (Accessed May 11, 2025).

Google Scholar

Reddy, S. (2024). Generative AI in healthcare: an implementation science informed translational path on application, integration and governance. Implement. Sci. 19:27. doi: 10.1186/s13012-024-01357-9

PubMed Abstract | Crossref Full Text | Google Scholar

Rubin, D. B. (1993). Statistical disclosure limitation. J. Off. Stat. 9, 461–468. Available online at: https://ecommons.cornell.edu/server/api/core/bitstreams/dd0b63ff-4494-4491-96ba-a69811563dee/content

Google Scholar

Ruczynski, K. (2024). Compare llms: a guide to finding the best large language models [online]. Word. Available online at: https://www.wordware.ai/blog/compare-llms-a-guide-to-finding-the-best-large-language-models (Accessed May 11, 2025).

Google Scholar

Rujas, M., Gomez, M., Del Moral Herranz, R., Fico, G., and Merino-Barbancho, B. (2025). Synthetic data generation in healthcare: a scoping review of reviews on domains, motivations, and future applications. Int. J. Med. Inform. 195:105763. doi: 10.1016/j.ijmedinf.2024.105763

PubMed Abstract | Crossref Full Text | Google Scholar

Sai, S., Gaur, A., Sai, R., Chamola, V., Guizani, M., and Rodrigues, J. J. P. C. (2024). Generative AI for transformative healthcare: a comprehensive study of emerging models, applications, case studies, and limitations. IEEE Access 12, 31078–31106. doi: 10.1109/ACCESS.2024.3367715

Crossref Full Text | Google Scholar

Scherr, R., Spina, A., Dao, A., Andalib, S., Halaseh, F. F., Blair, S., et al. (2025). Novel evaluation metric and quantified performance of chatgpt-4 patient management simulations for early clinical education: experimental study. JMIR Form. Res. 9:e66478. doi: 10.2196/66478

PubMed Abstract | Crossref Full Text | Google Scholar

Schick, T., and Schütze, H. (2021). "It’s not just size that matters: small language models are also few-shot learners", in: Proceedings of the 2021 conference of the north American chapter of the Association for Computational Linguistics (NAACL): Human language technologies, (Cambridge, MA: Association for Computational Linguistics), 2339–2352. doi: 10.18653/v1/2021.naacl-main.185

Crossref Full Text | Google Scholar

Schneider, J. (2025). Generative to agentic AI: survey, conceptualization, and challenges. arXiv [Preprint]. doi: 10.48550/arXiv.2504.18875

Crossref Full Text | Google Scholar

Schut, M. C., Luik, T. T., Vagliano, I., Rios, M., Helsper, C. W., van Asselt, K. M., et al. (2025). Artificial intelligence for early detection of lung cancer in gps' clinical notes: a retrospective observational cohort study. Br. J. Gen. Pract. 75, e316–e322. doi: 10.3399/BJGP.2023.0489

PubMed Abstract | Crossref Full Text | Google Scholar

Seinen, T. M., Kors, J. A., van mulligen, E. M., and Rijnbeek, P. R. (2024). Structured codes and free-text notes: measuring information complementarity in electronic health records. medRxiv [Preprint]. doi: 10.1101/2024.10.28.24316294

Crossref Full Text | Google Scholar

Sengupta, A., Goel, Y., and Chakraborty, T. (2025). How to upscale neural networks with scaling law? A survey and practical guidelines. arXiv [Preprint]. doi: 10.48550/arXiv.2502.12051

Crossref Full Text | Google Scholar

Shannon, C. E. (1948). A mathematical theory of communication. Bcl. Syst. Tech. J. 27, 379–423. doi: 10.1002/j.1538-7305.1948.tb01338.x

Crossref Full Text | Google Scholar

Shojaee, P., Mirzadeh, I., Alizadeh, K., Horton, M., Bengio, S., and Farajtabar, M. (2025). The illusion of thinking: understanding the strengths and limitations of reasoning models via the lens of problem complexity. arXiv [Preprint]. doi: 10.48550/arXiv.2506.06941

Crossref Full Text | Google Scholar

Singhal, K., Azizi, S., Tu, T., Mahdavi, S. S., Wei, J., Chung, H. W., et al. (2023). Large language models encode clinical knowledge. Nature 620, 172–180. doi: 10.1038/s41586-023-06291-2

PubMed Abstract | Crossref Full Text | Google Scholar

Smolyak, D., Bjarnadóttir, M. V., Crowley, K., and Agarwal, R. (2024). Large language models and synthetic health data: progress and prospects. JAMIA Open 7:ooae114. doi: 10.1093/jamiaopen/ooae114

PubMed Abstract | Crossref Full Text | Google Scholar

Steinkamp, J., Kantrowitz, J. J., and Airan-Javia, S. (2022). Prevalence and sources of duplicate information in the electronic medical record. JAMA Netw. Open 5:e2233348. doi: 10.1001/jamanetworkopen.2022.33348

PubMed Abstract | Crossref Full Text | Google Scholar

Suominen, H., Salanterä, S., Velupillai, S., Chapman, W. W., Savova, G., Elhadad, N., et al. (2013) Overview of the share/clef ehealth evaluation lab 2013 Information access evaluation. Multilinguality, multimodality, and visualization: 4th international conference of the CLEF initiative Heidelberg, Germany Springer, 212–231, doi: 10.1007/978-3-642-40802-1_17

Crossref Full Text | Google Scholar

Swart, N. M., Apeldoorn, A. T., Conijn, D., Meerhoff, G. A., and Ostelo, R. W. J. G. (2021). Kngf-richtlijn lage rugpijn en lumbosacraal radiculair syndroom [Online]. Amersfoort/ Utrecht: Koninklijk Nederlands Genootschap voor Fysiotherapie (KNGF) & Vereniging van Oefentherapeuten Cesar en Mensendieck. Available online at: https://vvocm.nl/Portals/2/kngf_richtlijn_lage_rugpijn_en_lrs_2021_verantwoording.pdf (Accessed May 15, 2025).

Google Scholar

Thandla, S. R., Armstrong, G. Q., Menon, A., Shah, A., Gueye, D. L., Harb, C., et al. (2024). Comparing new tools of artificial intelligence to the authentic intelligence of our global health students. BioData Min. 17:58. doi: 10.1186/s13040-024-00408-7

PubMed Abstract | Crossref Full Text | Google Scholar

Tischendorf, T., Hinsche, L., Hasseler, M., and Schaal, T. (2025). Genai in nursing and clinical practice: a rapid review of applications and challenges. J. Public Health (Berl.). doi: 10.1007/s10389-025-02523-z

Crossref Full Text | Google Scholar

Tuulos, V. (2022). Effective data science infrastructure: How to make data scientists productive. Shelter Island, NY: Manning.

Google Scholar

U.S.-National-Library-of-Medicine. (2024). Unified medical language system (UMLS) [Online]. Available online at: https://www.nlm.nih.gov/research/umls/index.html (Accessed May 15, 2025).

Google Scholar

Ulfsnes, R., Moe, N. B., Stray, V., and Skarpen, M. (2024). “Transforming software development with generative AI: empirical insights on collaboration and workflow” in Generative AI for effective software development. eds. A. Nguyen-Duc, P. Abrahamsson, and F. Khomh (Cham: Springer Nature Switzerland), 219–234.

Google Scholar

van der Willigen, R. F., Versnel, H., and van Opstal, A. J. (2024). Spectral-temporal processing of naturalistic sounds in monkeys and humans. J. Neurophysiol. 131, 38–63. doi: 10.1152/jn.00129.2023

PubMed Abstract | Crossref Full Text | Google Scholar

van Velzen, M., de Graaf-Waar, H. I., Ubert, T., van der Willigen, R. F., Muilwijk, L., Schmitt, M. A., et al. (2023). 21st century (clinical) decision support in nursing and allied healthcare. Developing a learning health system: a reasoned design of a theoretical framework. BMC Med. Inform. Decis. Mak. 23:279. doi: 10.1186/s12911-023-02372-4

Crossref Full Text | Google Scholar

Vasireddy, I., Sriveni, B., Prathyusha, K., and Kandi, A. (2024). Sentiment analysis of web media using hybrid model with t5 and gpt-4 models, in: 2nd international conference on recent trends in microelectronics, automation, computing and communications systems (ICMACC), (New York, NY: IEEE), 634–639. doi: 10.1109/ICMACC62921.2024.10894040

Crossref Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). “Attention is all you need”, in: Advances in Neural Information Processing Systems 30 (31st NeurIPS, 2017), eds. I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, and S. Vishwanathan, et al.: Curran Associates, Inc.), 15 pages. Available: https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf

Google Scholar

Verhulst, S., Zahuranec, A. J., and Chafetz, H. (2025). Moving toward the fair-r principles: Advancing AI-ready data. Available online at: SSRN.

Google Scholar

Walturn, (2025). Gpt-4.1 and the frontier of AI: Capabilities, improvements, and comparison to claude 3, gemini, mistral, and llama [Online]. Available online at: https://www.walturn.com/insights/gpt-4-1-and-the-frontier-of-ai-capabilities-improvements-and-comparison-to-claude-3-gemini-mistral-and-llama (Accessed May 13, 2025).

Google Scholar

Woisetschläger, H., Erben, A., Marino, B., Wang, S., Lane, N. D., Mayer, R., et al. (2024). Federated learning priorities under the European Union artificial intelligence act. arXiv [Preprint]. doi: 10.48550/arXiv.2402.05968

Crossref Full Text | Google Scholar

World Health Organization (2001). Icf: International classification of functioning, disability and health / world health organization. Geneva: World Health Organization.

Google Scholar

World Medical Association (2025). World medical association declaration of Helsinki: ethical principles for medical research involving human participants. JAMA 333, 71–74. doi: 10.1001/jama.2024.21972

PubMed Abstract | Crossref Full Text | Google Scholar

Xie, C., Cai, S., Wang, W., Li, P., Sang, Z., Yang, K., et al. (2025). Infir: crafting effective small language models and multimodal small language models in reasoning. arXiv [Preprint]. doi: 10.48550/arXiv.2502.11573

Crossref Full Text | Google Scholar

Xie, J., Chen, Z., Zhang, R., Wan, X., and Li, G. (2024). Large multimodal agents: a survey. arXiv [Preprint]. doi: 10.48550/arXiv.2402.15116

Crossref Full Text | Google Scholar

Yang, X., Xiao, Y., Liu, D., Deng, H., Huang, J., Zhou, Y., et al. (2025). Cross language transformation of free text into structured lobectomy surgical records from a multi center study. Sci. Rep. 15:15417. doi: 10.1038/s41598-025-97500-7

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, T., Kishore, V., Wu, F., Weinberger, K. Q., and Artzi, Y. (2019). Bertscore: evaluating text generation with bert. arXiv [Preprint]. doi: 10.48550/arXiv.1904.09675

Crossref Full Text | Google Scholar

Zhao, Z., Wallace, E., Feng, S., Klein, D., and Singh, S. (2021). "Calibrate before use: improving few-shot performance of language models", in: Proceedings of the 38th international conference on machine learning, eds. M. Marina and Z. Tong: PMLR, 12697––12706. Available online at: https://proceedings.mlr.press/v139/zhao21c.html

Google Scholar

Zheng, Y., Li, T., Huang, H., Zeng, T., Lu, J., Chu, C., et al. (2025). Are all prompt components value-neutral? Understanding the heterogeneous adversarial robustness of dissected prompt in large language models. arXiv [Preprint]. doi: 10.48550/arXiv.2508.01554

Crossref Full Text | Google Scholar

Keywords: healthcare, data synthesis, privacy, generative agents, linguistics, information theory, synthetic health data generation (SHDG), clinical natural language processing (NLP)

Citation: van Velzen M, van der Willigen RF, de Beer VJ, de Graaf-Waar HI, Janssen ERC, van Leeuwen S, van der Willigen MF, van der Willigen MJ, Renardus G, El Maaroufi R, Satimin SJ, Hartog LM, Hulsen T, van Meeteren NLU and Scheper MC (2025) Privacy-, linguistic-, and information-preserving synthesis of clinical documentation through generative agents. Front. Artif. Intell. 8:1644084. doi: 10.3389/frai.2025.1644084

Received: 09 June 2025; Accepted: 25 August 2025;
Published: 16 September 2025.

Edited by:

Tuan D. Pham, Queen Mary University of London, United Kingdom

Reviewed by:

Ricardo Valentim, Federal University of Rio Grande do Norte, Brazil
Mahesh Kumar Goyal, Google (United States), United States

Copyright © 2025 van Velzen, van der Willigen, de Beer, de Graaf-Waar, Janssen, van Leeuwen, van der Willigen, van der Willigen, Renardus, El Maaroufi, Satimin, Hartog, Hulsen, van Meeteren and Scheper. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Robert F. van der Willigen, ci5mLnZhbi5kZXIud2lsbGlnZW5AaHIubmw=

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.