ORIGINAL RESEARCH article
Front. Artif. Intell.
Sec. Technology and Law
This article is part of the Research TopicThe “Rule of AI”. Framing the future of artificial intelligence as a regulatory toolView all articles
Testing the Applicability of a Governance Checklist for High-Risk AI-Based Learning Outcome Assessment in Italian Universities Under the EU AI Act Annex III
Provisionally accepted- 1National Research Council (CNR), Institute for Educational Technology (ITD), Genova, Italy
- 2Universita degli Studi di Bari Aldo Moro Dipartimento di Giurisprudenza, Bari, Italy
- 3Department of Education Studies, Universita degli Studi di Bologna, Bologna, Italy
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Background: The EU AI Act classifies AI-based learning outcome assessment as high-risk (Annex III, point 3b), yet sector-specific frameworks for institutional self-assessment remain underdeveloped. This creates accountability gaps affecting student rights and educational equity, as institutions lack systematic tools to demonstrate that algorithmic assessment systems produce valid and fair outcomes. Methods: This exploratory study tests whether ALTAI's trustworthy AI requirements can be operationalized for educational assessment governance through the XAI-ED Consequential Assessment Framework, which integrates three educational evaluation theories (Messick's consequential validity, Kirkpatrick's four-level model, Stufflebeam's CIPP). Following pilot testing with three institutions, four independent coders applied a 27-item checklist to policy documents from 14 Italian universities (13% with formal AI policies plus one baseline case) using four-point ordinal scoring and structured consensus procedures. Results: Intercoder reliability analysis revealed substantial agreement (Fleiss's κ = 0.626, Krippendorff's α = 0.838), with higher alpha reflecting predominantly adjacent-level disagreements suitable for exploratory validation. Analysis of 14 universities reveals substantial governance heterogeneity among early adopters (Institutional Index: 0.00-60.32), with Technical Robustness and Safety showing lowest implementation (M=19.64, SD=21.08) and Societal Well-being highest coverage (M=52.38, SD=29.38). Documentation prioritizes aspirational statements over operational mechanisms, with only 13% of Italian institutions having adopted AI policies by September 2025. Discussion: The framework demonstrates feasibility for self-assessment but reveals critical misalignment: universities document aspirational commitments more readily than technical safeguards, with particularly weak capacity for validity testing and fairness monitoring. Findings suggest three interventions: (1) ministerial operational guidance translating EU AI Act requirements into educational contexts, (2) inter-institutional capacity-building addressing technical-pedagogical gaps, and (3) integration of AI governance indicators into national quality assurance systems to enable systematic accountability. The study contributes to understanding how educational evaluation theory can inform the translation of abstract trustworthy AI principles into outcome-focused institutional practices under high-risk classifications.
Keywords: Artificial intelligence in education, AI-based learning outcome assessment, Explainable AI, Altai, educational evaluation, Transparency, accountability, AI Governance
Received: 04 Oct 2025; Accepted: 25 Nov 2025.
Copyright: © 2025 Manganello, Nico, Ragusa and Boccuzzi. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Flavio Manganello
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
