Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Educ.

Sec. Assessment, Testing and Applied Measurement

Volume 10 - 2025 | doi: 10.3389/feduc.2025.1592037

This article is part of the Research TopicAI for Assessment, Testing and Applied MeasurementView all articles

Audit-Style Framework for Evaluating Bias in Large Language Models

Provisionally accepted
  • National Board of Medical Examiners, Philadelphia, United States

The final, formatted version of the article will be published soon.

One concern with AI systems is their potential to produce biased output. These biases can be difficult to detect due to the complex and often proprietary nature of the systems, which limits transparency. We propose an evaluation framework for assessing whether a system exhibits biased behavior. This evaluation consists of a series of tasks in which an AI system is instructed to select one of two students for an award based on their performance on a standardized assessment. The two students are implicitly associated with different demographic subgroups, and the evaluation is designed such that students from each subgroup perform equally well on average. In this way, any consistent preference for a particular subgroup can be attributed to bias in the system’s output.The proposed framework is illustrated using GPT-3.5 and GPT-4, with racial subgroups (Black/White) and an assessment composed of math items. In this demonstration, GPT-3.5 favored Black students over White students by a factor of approximately 2:1 (66.5%; 1,061 out of 1,596 non-equivocal choices). In contrast, GPT-4 showed a slight numerical preference for Black students (291 vs. 276; 51.3%), but this difference was not statistically significant (p = .557), indicating no detectable bias. These results suggest that the proposed audit is sensitive to differences in system bias in this context.

Keywords: Bias detection, Large Language Models (LLMs), educational assessment, AI fairness, Algorithmic bias, Audit-Style Evaluation

Received: 12 Mar 2025; Accepted: 27 Jun 2025.

Copyright: © 2025 Baldwin. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Peter Baldwin, National Board of Medical Examiners, Philadelphia, United States

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.