Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Public Health

Sec. Digital Public Health

This article is part of the Research TopicGenerative AI and Large Language Models in Medicine: Applications, Challenges, and OpportunitiesView all 4 articles

Effects of Education Level on Natural Language Processing in Cardiovascular Health Communication

Provisionally accepted
Stanley  JosephStanley Joseph1*Ashna  BhardwajAshna Bhardwaj1Justin  SkariahJustin Skariah1Ishan  AggarwalIshan Aggarwal1Varunil  ShahVarunil Shah2Ryan  A HarrisRyan A Harris1*
  • 1Augusta University Medical College of Georgia, Augusta, United States
  • 2Case Western Reserve University School of Medicine, Cleveland, United States

The final, formatted version of the article will be published soon.

ABSTRACT Cardiovascular disease (CVD) remains the leading cause of mortality worldwide, underscoring the importance of accessible health communication. Artificial intelligence (AI) tools such as ChatGPT and MediSearch have potential to bridge knowledge gaps, but their effectiveness depends on both accuracy and readability. This study evaluated how natural language processing (NLP) models respond to CVD-related questions across different education levels. Thirty-five frequently asked questions from reputable sources were reformatted into prompts representing lower secondary, higher secondary, and college graduate levels, and entered into ChatGPT Free (GPT-4o mini), ChatGPT Premium (GPT-4o), and MediSearch (v1.1.4). Readability was assessed using Flesch-Kincaid Ease and Grade Level scores, and response similarity was evaluated with BERT-based cosine similarity. Statistical analyses included ANOVA, Kruskal-Wallis, and Pearson correlation. Readability decreased significantly with increasing education level across all models (p<0.001). ChatGPT Free responses were more readable than MediSearch (p<0.001), while ChatGPT Free and Premium demonstrated higher similarity to each other than to MediSearch. ChatGPT Premium explained the greatest variance in readability (r=0.350; p<0.001), suggesting stronger adaptability to user education levels compared to ChatGPT Free (r=0.530; p<0.001) and MediSearch (r=0.227; p<0.001). These findings indicate that while NLP models adjust readability by education level, output complexity often exceeds average literacy, highlighting the need for refinement to optimize AI-driven patient education.

Keywords: cardiovascular disease, artificial intelligence, Natural Language Processing, Health Communication, readability, Patient Education, Large language models

Received: 20 Aug 2025; Accepted: 28 Oct 2025.

Copyright: © 2025 Joseph, Bhardwaj, Skariah, Aggarwal, Shah and Harris. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence:
Stanley Joseph, stjoseph@augusta.edu
Ryan A Harris, ryharris@augusta.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.