Your new experience awaits. Try the new design now and help us make it even better

SYSTEMATIC REVIEW article

Front. Comput. Sci., 06 January 2026

Sec. Digital Education

Volume 7 - 2025 | https://doi.org/10.3389/fcomp.2025.1721093

This article is part of the Research TopicEnhancing Learning with Online Educational Videos in the Web 2.0 Era: Learner Engagement, Learning Processes and OutcomesView all 4 articles

A rapid review of using AI-generated instructional videos in higher education

Tran Trieu Hai,Tran Trieu Hai1,2Duong Thi Thuy MaiDuong Thi Thuy Mai1Nguyen Van Hanh
Nguyen Van Hanh1*
  • 1Faculty of Education, Hanoi University of Science and Technology, Hanoi, Vietnam
  • 2Faculty of General Education, Hanoi Open University, Hanoi, Vietnam

Introduction: Generative artificial intelligence (AI) has enabled the rapid emergence of AI-generated instructional videos (AIGIVs) as a new form of learning material in higher education. However, evidence on how they are produced, applied, and the reported benefits and risks remains fragmented, highlighting the need for a systematic synthesis.

Methods: This study conducted a rapid review following PRISMA principles. Studies published from 2023 onward were searched on the Web of Science, Scopus, IEEE Xplore, and Google Scholar. Fifteen eligible studies were analyzed using qualitative content analysis and thematic synthesis.

Results: Two production modes were identified: fully AI-based video generation (e.g., Sora, HeyGen, Veo) and AI-assisted human-made production (e.g., DALL·E, ChatGPT). Pedagogical applications included using AIGIVs as instructional alternatives and as tools for reflective pedagogy, particularly ethical and critical reflection. Benefits included efficiency and scalability, improved accessibility and personalization, and enhanced emotional engagement and memory. Risks involved ethical concerns, technical limitations, and inauthentic or unreliable content.

Discussion: AIGIVs show strong potential for higher education, but their value depends on instructional design, human oversight, and responsible governance.

1 Introduction

The COVID-19 pandemic has created a strong acceleration of digital transformation, forcing higher education institutions to rapidly shift to more flexible teaching and learning formats (Lan et al., 2020; Long and Van Hanh, 2020; Tonbuloğlu and Tonbuloğlu, 2023; Tuyet et al., 2020). E-learning and blended learning have emerged as urgent solutions, while also reshaping the long-term development strategies of higher education institutions (Bozkurt and Sharma, 2020). In this form of learning, instructional videos become the “heart” of the digital learning ecosystem (Guo et al., 2014). Instructional videos are present in most modern learning management systems, from regular university courses to distance learning programs. In e-learning, videos often serve as the primary learning material for content delivery, replacing face-to-face lectures. In blended learning, video lectures serve as the core of online learning activities, previewed by students at home before engaging in face-to-face classroom practice and discussion (Bishop and Verleger, 2013).

Video-based learning refers to the process of acquiring defined knowledge, competencies, and skills through the systematic use of video resources (Chen and Feng, 2023). In digital environments, video-based learning enables learners to engage with content through visual, auditory, and interactive means (Guo et al., 2014). Another key advantage of video is its flexibility: learners can rewatch content multiple times, revisit difficult segments, adjust playback speed, or learn at times that best fit their personal schedules (El-Ariss et al., 2021; Xie et al., 2017; Zhang et al., 2006). Furthermore, with the integration of technologies such as adaptive learning and personalized content delivery, learners can access videos tailored to their learning pace, cognitive level, or personal preferences (Mo et al., 2022; Sanal Kumar and Thandeeswaran, 2025). For example, interactive tools like Edpuzzle and H5P allow learners to answer questions, complete quick assessments, provide feedback, or receive immediate correction while watching the video (Di Cesare et al., 2021). A study investigating the integration of H5P interactive video content into a learning management system found that university students in e-learning courses reported significantly higher satisfaction than students in other environments (Mir et al., 2021). Additionally, instructional video, particularly the talking-head format, where the instructor appears on screen, can enhance feelings of closeness, connection, and help establish “teaching presence,” a critical element of the community of inquiry model (Chen and Feng, 2023). The instructor’s voice and visual presence help learners feel as though they are being “taught” rather than studying alone (Lazarevic, 2011).

Unlike traditional educational videos which are typically produced by human instructors and instructional design experts (Netland et al., 2025), the emergence of generative artificial intelligence (AI) technologies such as Synthesia, HeyGen, Sora, and Veo has given rise to a new form of instructional material: AI-generated instructional videos (AIGIVs). These videos are defined as instructional materials in which part or all of the content, visuals, voice, or presenter is automatically generated by AI technologies, including natural language processing (NLP), text-to-speech, AI avatars, and text-to-video (Izani et al., 2025; Shu et al., 2025). The benefits of using AI in video production lie in the speed of production, low cost, and potential for large-scale personalization (Netland et al., 2025). AI tools for video creation can democratize production, making video production more accessible to individuals and smaller organizations, especially those who may not have the resources for traditional video production methods (Pellas, 2023).

As universities increasingly experiment with integrating AI into teaching and learning, the emerging presence of AIGIVs has added to the excitement of innovative educators. Early studies have already begun to document the use of AIGIVs, signaling an urgent need to understand their broader implications. Considering this fast-moving trend, it is essential to take stock of the current landscape: What benefits have been observed so far? What risks or limitations are emerging? And what do early studies suggest about the role these videos might play in future instructional design? Given the novelty and potential impact of this trend in education, a rapid review of the existing literature is both timely and necessary. Such a review can help synthesize current evidence, clarify the opportunities and challenges of AIGIVs, and identify gaps that require further investigation. It also lays an important foundation for making informed pedagogical decisions, designing responsible organizational strategies, and shaping future research in the growing domain of AIGIVs.

Therefore, this study aims to conduct a rapid review of existing literature to synthesize emerging evidence on the use of AIGIVs in higher education. The temporal boundary of 2023–2025 was set to capture the post-ChatGPT era, during which generative AI technologies including ChatGPT (launched November 2022), DALL·E 2, HeyGen, Sora, Synthesia, and Veo 2 began to enable fully automated or AI-assisted video production. This review addresses the following research questions:

Research question 1 (RQ1): What modes and technologies have been employed to create AIGIVs for higher education, and how do these approaches differ in terms of automation and human involvement?

Research question 2 (RQ2): How have AIGIVs been integrated into teaching and learning practices, and what pedagogical functions do they serve?

Research question 3 (RQ3): What benefits and risks have been reported regarding the use of AIGIVs in higher education?

2 Methodology

2.1 Research design

This study used a rapid review design to provide a timely synthesis of the emerging literature on AIGIVs in higher education. Rapid reviews are simplified forms of systematic reviews that retain the core principles of transparency and rigor, to provide evidence more quickly for policy and practice. In line with this approach, the review followed the key steps recommended in PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses), including explicit eligibility criteria, a multi-database search strategy, systematic screening, and transparent reporting of study selection. To improve methodological clarity, the research questions were also framed using a PICO structure:

• Population (P): Students and educators in higher education, as well as adult learners in professional and societal sectors where AIGIVs are explicitly used as instructional materials.

• Intervention (I): The intervention focused on AIGIVs, defined as videos in which part or all of the content, visuals, voice, or presenter is generated by AI tools (e.g., Sora, HeyGen, DALL·E 2, AI avatars, text-to-video tools).

• Comparator (C): AIGIVs were compared with instructor-made videos or AIGIVs were examined as a standalone intervention.

• Outcomes (O): Reported learning outcomes, learner and instructor perceptions, ethical and epistemic concerns, and other pedagogical or organizational implications.

This rapid review does not aim to provide an exhaustive, comprehensive map of the field but rather to synthesize the most recent, peer-reviewed evidence (2023 onwards) on how AIGIVs are being produced and used, and what benefits and risks have been identified so far.

2.2 Search strategy

The literature search was conducted across four major academic databases: Web of Science, Scopus, and IEEE Xplore. The Google Scholar database was used to search for additional studies in the first 200–300 results displayed (Haddaway et al., 2015).

The online search was conducted in September 2025, focusing on literature published from 2023 onwards, which marks a period of widespread use of generative AI tools such as ChatGPT and Sora. The following Boolean search string was applied:

(“text to video” OR “text-to-video” OR “AI video generation” OR “AI-generated videos” OR “AI-generated learning videos” OR “AI-generated teaching videos” OR “AI-generated instructional videos” OR “AI-generated avatars”) AND (“education”)

The search was limited to each database as follows:

• Web of Science: Articles, proceeding papers, and reviews.

• Scopus: Document type (conference paper, article, and review), and language (English).

• IEEE Xplore: Journal articles and conference papers.

• Google Scholar: Used to identify additional relevant articles not indexed in the above databases.

2.3 Inclusion and exclusion criteria

Studies were included if they met all the following conditions:

• Explicitly investigated the use or pedagogical application of AIGIVs in educational, training, or professional learning settings;

• Addressed higher education or closely related contexts (e.g., adult learning, or professional education).

• Reported empirical data, conceptual analyzes, or design-based implementations related to teaching, learning, or assessment using AIGIVs.

Studies were excluded during full-text screening if they:

• Focused solely on technical video generation algorithms/ models without educational applications;

• Examined AI in education in general but did not involve AIGIVs as part of instructional design or delivery;

• Investigated traditional instructor-made videos or generic video-based learning lacking AI-generated components; were not available in full text.

2.4 Screening process

Figure 1 presents a PRISMA flow diagram summarizing the screening process. The initial search yielded a total of 118 studies across four databases: 30 from Web of Science, 52 from Scopus, 31 from IEEE Xplore, and 5 additional studies manually identified through Google Scholar. After removing 12 duplicates, 106 studies remained for title-and-abstract screening.

Figure 1
PRISMA flow diagram showing study selection for a rapid review of AI-generated instructional videos in higher education. Records identified from databases total 118 studies. After removing 12 duplicates, 106 records remain for title/abstract screening. Many are excluded for not involving AI-generated videos (15), being technical-only (48), or outside higher education/adult learning (8). Thirty-five full texts are assessed; exclusions include technical-only (8), not using AIGIVs (2), traditional video-only (1), or irrelevant results (9). Final included studies: 15.

Figure 1. PRISMA flow diagram.

At the title-and-abstract screening stage, studies were excluded for several reasons: the paper did not mention or involve AI-generated videos (n = 15); the study focused solely on the technical development of AI video generation tools (n = 48); or the study was outside the higher education or adult learning context (n = 8).

A total of 35 full-text studies were then assessed in detail. At the full-text screening stage, studies were excluded when they: focused solely on technical algorithms or models for video generation without educational applications (n = 8); addressed AI in education but did not use AIGIVs in instructional design or delivery (n = 2); examined general video-based learning or traditional instructor-made videos without AI-generated components (n = 1); or did not report results relevant to the research questions (n = 9). Ultimately, 15 studies met all inclusion criteria and were included in the final synthesis.

2.5 Data extraction and analysis

Key information from the 15 included studies was systematically extracted into a structured Excel matrix. For each study, the following fields were recorded: publication year, country, research design, education or training level, and PICO components. Characteristics of the included studies are summarized in Table 1.

Table 1
www.frontiersin.org

Table 1. Characteristics of included studies.

As shown in Table 1, reviewed studies used a variety of research designs, including exploratory analyzes, mixed-methods, surveys, qualitative interviews, experimental designs, and narrative reviews. The geographical distribution includes contributions from countries such as Nigeria, the United Kingdom, Greece, Australia, the United States, Canada, China, Switzerland, the UAE, Austria, Italy, and Saudi Arabia, reflecting a global interest in AIGIVs. A variety of AI tools were reported across the studies, with Sora emerging as the most frequently examined platform. Other tools included Veo 2, HeyGen, ChatGPT, and DALL·E 2. The included studies addressed a range of aims, from exploring the technical capabilities of AI tools like Sora and HeyGen to evaluating pedagogical effectiveness and learner perceptions. Several studies also examined ethical concerns and public attitudes, reflecting the multifaceted perspectives on the use of AI-generated video in education.

To answer the research questions, a qualitative content analysis was conducted to systematically interpret patterns and meanings within the reviewed studies. The analysis followed three iterative stages. First, all extracted findings were segmented into 45 meaning units and assigned descriptive codes that represented key ideas (e.g., automating instructional video creation with Sora, producing avatar-based instructional videos via HeyGen …). Second, similar codes were grouped into eight conceptual categories that reflected broader aspects such as technical limitations, and ethical concerns. Finally, these categories were summarized into four major themes: (1) Modes and technologies of AI-based instructional video generation, (2) Applying AIGIVs for instructional and reflective pedagogies, (3) Reported benefits of integrating AIGIVs into education, and (4) Reported risks and challenges of integrating AIGIVs into education. The results were detailed in Table 2.

Table 2
www.frontiersin.org

Table 2. Codes, categories, and themes identified from the reviewed studies on AIGIVs.

2.6 Reliability

As this review was conducted by a single author, reliability was enhanced through peer verification with two colleagues in educational technology. One researcher independently participated in reading and extracting textual segments from the included studies. Agreement between the author and this researcher was calculated for the extraction of 46 meaning units, reaching 41 agreements (≈89%), which indicates a high level of consistency. For the remaining disagreements, a third expert in educational technology was invited to join a consensus discussion. During this session, the three participants jointly reviewed all extracted segments and reached a final consensus on 45 meaning units. We also confirmed the allocation of each meaning unit to its corresponding codes and refined the grouping of categories and overarching themes.

3 Findings

3.1 Theme 1. Modes and technologies of AI-based instructional video generation

Figure 2 illustrates the distribution of AI tools used in instructional video production (N = 15 coded instances). Fully AI-generated videos accounted for 73.4% of the reviewed evidence, while AI-assisted human-made video production received less attention. Sora was most frequently studied (46.7%), followed by HeyGen (20.0%) and DALL·E 2 (13.3%), and ChatGPT (13.3%), while Veo 2 (6.7%) appeared less often. This distribution indicates a strong trend toward fully automated video-generation approaches.

Figure 2
Chart summarizing the AI tools used in instructional video production across the reviewed evidence (N = 15 coded instances). Fully AI-generated video approaches dominate at 73.4%, while AI-assisted human-made production is less common (26.6%). Among specific tools, Sora appears most frequently (46.7%), followed by HeyGen (20.0%). DALL·E 2 and ChatGPT each account for 13.3%, and Veo 2 appears least often (6.7%). The visualization emphasizes a strong research focus on fully automated, text-to-video or avatar-based generation platforms.

Figure 2. AI technologies used in instructional video production.

3.1.1 Category 1.1: generating fully AI-based instructional videos

Eleven of the fifteen studies analyzed focused on automating video creation through generative AI models capable of producing entire instructional videos with minimal human input. The most frequently mentioned system was Sora, a text-to-video model used in contexts ranging from storytelling to medical training simulations (Adetayo et al., 2024; Dağcı et al., 2025; Mogavi et al., 2024; Mohamed and Lucke-Wold, 2024; Temsah et al., 2025; Waisberg et al., 2024; Zhou et al., 2024). Similarly, Veo 2 was employed in patient education contexts to generate health-related instructional videos (Temsah et al., 2025). HeyGen was also used in a fully automated mode in three studies to create videos with AI avatars and voice synthesis from script-based inputs (Netland et al., 2025; Pellas, 2023; Struger et al., 2025). Overall, these tools represent a new wave of generative AI applications focused on efficiency and automation.

3.1.2 Category 1.2: co-creating instructional videos with AI assistance

In contrast, a smaller group of studies explored AI-assisted video production, where educators remain the primary designers while AI systems serve as creative partners. For example, DALL·E 2 was used to generate visual components integrated into instructor-produced videos (Leiker et al., 2023; Netland et al., 2025). ChatGPT, combined with multimodal tools such as PowerPoint or/and SADTalker, supported educators in co-creating instructional videos, or drafting scripts (Leiker et al., 2023; Xu et al., 2025). These examples illustrate that AI assistance can enhance creative efficiency while preserving pedagogical control, blending human expertise in content and pedagogy with machine-generated visuals and narration.

3.2 Theme 2. Applying AIGIVs for instructional and reflective pedagogies

Theme 2 highlights two major pedagogical directions of AI-generated instructional videos (AIGIVs): (1) adopting AIGIVs as instructional alternatives and (2) integrating AIGIVs into AI-enhanced reflective pedagogies.

3.2.1 Category 2.1: adopting AIGIVs as instructional alternatives

Five studies compared AI-generated videos with instructor-made videos to evaluate their pedagogical effectiveness (Leiker et al., 2023; Miranda and Vegliante, 2025; Netland et al., 2025; Struger et al., 2025; Xu et al., 2025). Across these experiments, AIGIVs yielded learning outcomes comparable to those achieved through instructor-produced videos, suggesting that they can feasibly act as replacements in certain instructional contexts. For example, in management courses, AI-generated avatars and narrators resulted in learning outcomes comparable to human-generated videos for undergraduate students (Netland et al., 2025). Additionally, research on student perceptions revealed generally favorable attitudes toward AI-generated avatars (Pellas, 2023; Vallis et al., 2024).

3.2.2 Category 2.2: AI-enhanced reflective pedagogy

Beyond direct content delivery, three studies positioned AIGIVs as catalysts for ethical and critical reflection. Two studies described using AI avatars to encourage ethical reflection and moral reasoning in social science and business ethics courses (Pellas, 2023; Vallis et al., 2024). In one case, the authors redesigned an entire ethics curriculum around AI avatar-based video scenarios, creating opportunities for students to explore complex dilemmas (Vallis et al., 2024). Subsequently, 10 interviewed students felt that AIGIVs were suitable for teaching, even preferable in some cases to human instructors if properly designed (Vallis et al., 2024). Another study mentioned the use of Sora as a powerful tool that provides opportunities for creative and reflective learning through immersive storytelling videos (Adetayo et al., 2024).

3.3 Theme 3. Reported benefits of integrating AIGIVs into education

Theme 3 synthesizes the reported benefits of integrating AIGIVs into educational settings. As summarized in Table 2, three benefit-oriented categories emerged: (1) improving efficiency and scalability of video production, (2) enhancing accessibility and personalization in learning, and (3) enhancing emotional engagement and memory retention.

3.3.1 Category 3.1: improving efficiency and scalability of video production

Across the four related codes, AIGIVs were consistently associated with greater efficiency and scalability in video production. Studies on Sora and related AI video tools reported that educators and institutions could produce instructional videos more quickly and at lower cost compared with traditional recording workflows (Adetayo et al., 2024; Mogavi et al., 2024; Temsah et al., 2025). Furthermore, public and professional commentaries highlighted that AI video platforms can democratize educational video production, enabling smaller institutions, individual educators, and even learners to participate in content creation without substantial financial or technical resources (Mogavi et al., 2024). In general, these findings suggest that AIGIVs can support more sustainable and scalable models of video-based education, especially in contexts where production capacity has traditionally been constrained.

3.3.2 Category 3.2: enhancing accessibility and personalization in learning

This category, formed by two codes, emphasizes how AIGIVs can improve accessibility and personalization. For example, in multilingual higher education settings, AI-generated virtual speakers were used to provide content in multiple languages, making e-learning materials more accessible to diverse student populations (Miranda and Vegliante, 2025). These videos allowed learners to select language options according to their preferences, thereby reducing linguistic barriers. Moreover, AI-generated videos were reported to support personalized learning experiences, for example by adapting pacing and presentation style (Miranda and Vegliante, 2025). Although the current evidence base is small, this initial evidence indicates that AIGIVs can contribute to more inclusive and learner-centered environments when combined with thoughtful instructional design.

3.3.3 Category 3.3: enhancing emotional engagement and memory retention

The third benefit category focuses on emotional engagement and memory retention, drawing on three codes (Struger et al., 2025; Xu et al., 2025). For example, experimental findings suggested that AI-generated teaching avatars can capture learners’ attention, evoke stronger emotional responses, and support better recall of key ideas compared with videos with real instructors (Struger et al., 2025). In language and higher education contexts, students reported that AIGIVs were memorable and engaging, which in turn supported their ability to remember and apply content (Xu et al., 2025).

3.4 Theme 4. Reported risks and challenges of integrating AIGIVs into education

Theme 4 synthesizes the risks and challenges associated with integrating AIGIVs into higher education. As summarized in Table 2, 11 risk-related codes were grouped into three categories: ethical concerns, technical and design limitations, and epistemic reliability and authenticity. When comparing production modes, all these risks and challenges were associated with fully AI-generated video.

3.4.1 Category 4.1: ethical concerns

Three studies have highlighted ethical concerns surrounding the creation and use of AI-generated videos. For example, an analysis of 292 public comments on Sora revealed worries that AIGIVs could threaten human autonomy, data privacy, and copyright, such as enabling unauthorized replication of individuals’ images or obscuring who controls the content (Zhou et al., 2024). Similarly, further comment analyzes pointed to misinformation and bias risks, with users concerned that AI-generated videos might be deployed to spread distorted narratives (Mogavi et al., 2024). Additionally, deepfake-related ethical risks were raised in medical and public-education contexts, where highly realistic but synthetic videos could be used to fabricate events or statements (Temsah et al., 2025). These concerns highlight the need for clear labeling and increased AI literacy (Zhou et al., 2024).

3.4.2 Category 4.2: technical limitations

AIGIVs still have technical issues that can affect their effectiveness in education. For example, existing research shows that Sora still has significant limitations when visualizing content, such as physically implausible motion generation, unnatural object morphing, inaccurate physical interactions, and abnormal behavior presentation when generating videos in the field of neurosurgery (Mohamed and Lucke-Wold, 2024). Similarly, videos generated by Sora also contained errors in anatomical content and in common structures such as fingers or teeth (Waisberg et al., 2024). These findings show that integrating AIGIVs into curricula requires rigorous validation of the content quality of AIGIVs.

3.4.3 Category 4.3: inauthentic and unreliable content

Finally, some researchers have expressed concerns about whether AIGIVs can be trusted. One important concern noted is the blurred boundaries between authentic and artificial content (Dağcı et al., 2025; Zhou et al., 2024). In educational settings, this ambiguity may lead learners to question whether the instructor is a real expert or an AI simulation. Additionally, other studies have reported concerns about information accuracy in AIGIVs (Mogavi et al., 2024; Struger et al., 2025; Waisberg et al., 2024). This poses the risk of spreading misinformation.

4 Discussion

This rapid review synthesized emerging evidence on how AIGIVs are produced and used in higher education, and how their adoption gives rise to both pedagogical opportunities and critical risks. Figure 3 presents an integrative flow model that links the four themes identified in the findings: (1) modes and technologies of AI-based instructional video generation, (2) applying AIGIVs for instructional and reflective pedagogies, (3) reported benefits of integrating AIGIVs into education, and (4) reported risks and challenges. Rather than viewing technologies, pedagogical uses, and outcomes as separate elements, the model emphasizes their sequential and interconnected nature.

Figure 3
Conceptual flow model showing how AI-generated instructional videos (AIGIVs) move from production to educational impacts. On the left, two production modes are distinguished: fully AI-based video generation (e.g., Sora, HeyGen, Veo 2) and AI-assisted human-made production (e.g., DALL·E 2, ChatGPT). The middle section shows pedagogical applications, including using AIGIVs as instructional alternatives and for reflective pedagogy (ethical/critical reflection). On the right, outcomes split into benefits (efficiency/scalability, accessibility/personalization, emotional engagement/memory) and risks (ethical concerns, technical limitations, inauthentic or unreliable content).

Figure 3. Flow model of integrating AIGIVs into education.

On the left side of the model, AI technologies are grouped into two distinct production modes: fully AI-based instructional video generation (e.g., Sora, HeyGen, Veo 2) and AI-assisted human-made video production (e.g., DALL·E 2, ChatGPT). Fully automated tools offer speed and scalability, enabling low-cost video creation with minimal human effort (Adetayo et al., 2024; Temsah et al., 2025). At the same time, this mode also concentrates most of the reported risks, including ethical concerns, technical limitations, and inauthentic and unreliable content. In contrast, AI-assisted workflows retain a stronger role for educators in shaping content and context (Leiker et al., 2023). This hybrid mode appears less frequently in the current literature but is consistently associated with more fine-grained control over quality, and alignment with learning goals.

The central part of the model concerns how AIGIVs are pedagogically applied. The first approach involves adopting AIGIVs as instructional alternatives, where AI-generated videos replace instructor-made recordings in lectures or explanatory segments. Across multiple experiments, such substitutions did not lead to significant differences in short-term learning performance when compared with human-made videos (Leiker et al., 2023; Miranda and Vegliante, 2025; Netland et al., 2025; Struger et al., 2025; Xu et al., 2025). These findings suggest that AIGIVs can function as viable substitutes for conventional video lectures, offering comparable cognitive outcomes while reducing production effort. However, several studies also reported that learners still value the social and emotional presence of human instructors (Netland et al., 2025), which indicates that replacement should be considered carefully. The second approach emphasizes AI-enhanced reflective pedagogy. AIGIVs are not treated merely as vehicles for content delivery but as prompts for ethical reasoning, discussion, and metacognitive engagement. For example, AI avatar-based instructional videos were used to present complex dilemmas and stakeholder perspectives in business ethics and social science courses, encouraging students to analyze, debate, and justify their decisions (Pellas, 2023; Vallis et al., 2024). This approach highlights the promising applications of AIGIV for students’ higher-order learning.

The right side of the model captures both the benefits and risks that emerge from pedagogical applications of AIGIVs. On the benefit side, the reviewed studies converge on three main contributions of AIGIVs. First, AIGIVs can improve efficiency and scalability of video production, enabling institutions and individual educators to generate instructional content more quickly and at lower cost, while potentially democratizing participation in content creation (Adetayo et al., 2024; Mogavi et al., 2024; Temsah et al., 2025). Second, AIGIVs can enhance accessibility and personalization, particularly in multilingual learning environments, where AI-generated virtual speakers and adaptive videos can be tailored to learners’ linguistic needs (Miranda and Vegliante, 2025). In other examples, AI avatar-based instructional videos can boost emotional engagement and memory retention of students (Struger et al., 2025; Xu et al., 2025). In addition, this study also notes a substantial cluster of risks and challenges that must be addressed for AIGIVs to be used responsibly in education, including ethical concerns (Mogavi et al., 2024; Temsah et al., 2025; Zhou et al., 2024), technical limitations (Mohamed and Lucke-Wold, 2024; Waisberg et al., 2024), and inauthentic and unreliable content (Dağcı et al., 2025; Mogavi et al., 2024; Struger et al., 2025; Waisberg et al., 2024; Zhou et al., 2024). Thus, the flow model reinforces that benefits and risks are structurally linked to how AIGIVs are produced and pedagogically deployed.

Beyond mapping relationships among themes, the model also has practical implications for instructional design. Viewed through the lens of established instructional design frameworks such as ADDIE or ASSURE, the proposed flow model should be understood primarily as a conceptual design idea. It illuminates the conceptual space where technological, pedagogical, and ethical dimensions of AIGIVs intersect. It opens a reflective dialogue about how AIGIVs can be intentionally designed, critically interpreted, and responsibly embedded within diverse educational contexts. For educators, the model provides a structured yet flexible way to promote reflective decision-making across all stages of instructional video design. For researchers, it serves as a generative framework for exploring how AI transforms instructional video design thinking, and inviting empirical inquiry into how AIGIVs evolve in real learning environments. For policymakers, it highlights the necessity of transparent standards, institutional safeguards, and ethical governance frameworks that ensure generative AI effectively serves educational purposes.

Finally, the patterns captured in the model should be interpreted in light of the rapid review design and the still nascent evidence base. The synthesis is based on 15 studies published between 2023 and 2025, spanning diverse disciplines, countries, and methodological approaches. This breadth is a strength in capturing the early landscapes of the field, but it also means that the conclusions are indicative rather than definitive and cannot be generalized to all higher education contexts. As research on AIGIVs matures, future studies should test and refine the proposed flow model, explore long-term learning effects, and consider institutional policies regarding the integration of AIGIVs into instructional practices.

5 Conclusion

This rapid review has synthesized emerging research on the use of AIGIVs in higher education, highlighting three interconnected dimensions: the technologies used in video production, the pedagogical applications of AIGIVs, and the resulting educational benefits and risks. Findings suggest that while AIGIVs offer promising advantages in terms of scalability, efficiency, and innovation, their effectiveness depends on how they are integrated into instructional design and educational contexts. Technical limitations, ethical concerns, and epistemic risks must be carefully addressed to ensure responsible implementation. The proposed flow model illustrates the progression from technological deployment to pedagogical use and educational impact, reinforcing the need for a system-thinking approach. As AIGIVs continue to evolve, educators, researchers, and policymakers must collaborate to develop robust frameworks that maximize pedagogical value while safeguarding educational integrity.

5.1 Recommendations for educational practice and future research

Although research on AIGIVs is rapidly emerging, most existing studies remain preliminary, focusing on short-term outcomes and initial user perceptions. Based on the thematic synthesis of current literature, several key recommendations can be made for educational practice and future research:

5.1.1 Move beyond performance comparisons to explore pedagogical value and deep learner experiences

Much of the current research focuses narrowly on comparing AIGIVs with instructor-made videos in terms of short-term learning outcomes such as test scores or self-reported satisfaction from students (Leiker et al., 2023; Netland et al., 2025). While these studies offer initial insights, they do not capture the deeper pedagogical potential or affective–cognitive impact of AIGIVs. Future research should shift toward investigating students’ learning mechanisms activated when engaging with AIGIVs, and examine the role of their emotions, motivation, and reflection in the AIGIVs-based learning process.

5.1.2 Develop pedagogical design frameworks that meaningfully integrate AIGIVs

Studies have shown that the effectiveness of AIGIVs is highly dependent on how well they align with learning objectives and student characteristics (Vallis et al., 2024; Waisberg et al., 2024). Therefore, there is a need to create theoretical frameworks and instructional models that define the role of AIGIVs, whether as replacements, supplements, or reflective triggers.

5.1.3 Enhance digital and AI literacy for both learners and educators

Learners’ prior experience with technology significantly shapes their attitudes toward AIGIVs (Pellas, 2023). Likewise, educators who lack foundational understanding of AI may struggle to evaluate the quality, control the accuracy, or strategically use AIGIVs in teaching (Pellas, 2023). Therefore, institutions should offer professional development programs in AI literacy for educators, focusing on how to use, assess, and integrate AI tools in teaching. At the same time, students should be equipped with critical digital competencies to evaluate AI-generated content independently and responsibly.

5.1.4 Establish policies and ethical safeguards for AIGIVs deployment

Several studies have raised concerns regarding the ethical implications of AIGIVs, including deepfakes, misinformation, and blurred boundaries between authentic and artificial content (Mogavi et al., 2024; Temsah et al., 2025; Zhou et al., 2024). In educational contexts, where academic integrity is paramount, the following policy measures are recommended:

• Mandatory labeling of AI-generated content in instructional materials.

• Development of quality standards for AIGIVs in educational settings.

In summary, while AIGIVs offer significant promise for transforming the production of digital learning materials with scalability and cost-effectiveness, realizing its benefits while mitigating risks requires a system-level approach that incorporates not just technological considerations, but also pedagogical strategies, policy frameworks, and human capacity building.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

TH: Methodology, Formal analysis, Resources, Writing – original draft. DM: Conceptualization, Methodology, Formal analysis, Writing – original draft. NH: Funding acquisition, Validation, Project administration, Supervision, Conceptualization, Formal analysis, Writing – review & editing, Methodology, Writing – original draft, Data curation, Visualization, Investigation, Software, Resources.

Funding

The author(s) declared that financial support was not received for this work and/or its publication.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that Generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Adetayo, A. J., Enamudu, A. I., Lawal, F. M., and Odunewu, A. O. (2024). From text to video with AI: the rise and potential of Sora in education and libraries. Libr. Hi Tech News. doi: 10.1108/LHTN-02-2024-0028,

PubMed Abstract | Crossref Full Text | Google Scholar

Bishop, J., and Verleger, M. A.. (2013). The flipped classroom: a survey of the research. 2013 ASEE Annual Conference & Exposition, 23–1200.

Google Scholar

Bozkurt, A., and Sharma, R. C. (2020). Emergency remote teaching in a time of global crisis due to coronavirus pandemic. Asian J. Distance Educ. 15, i–vi. doi: 10.5281/zenodo.3778083

Crossref Full Text | Google Scholar

Chen, X., and Feng, S. (2023). Exploring the relationships between social presence and teaching presence in online video-based learning. J. Comput. Assist. Learn. 39, 1769–1785. doi: 10.1111/jcal.12843

Crossref Full Text | Google Scholar

Dağcı, M., Dost, A., and Çam, F. (2025). Potential impacts of AI-generated videos on nursing care. Nurse Educ. 50, E254–E258. doi: 10.1097/NNE.0000000000001880,

PubMed Abstract | Crossref Full Text | Google Scholar

Cesare, D. M. D., Kaczorowski, T., and Hashey, A. (2021). A piece of the (Ed)puzzle: using the Edpuzzle interactive video platform to facilitate explicit instruction. J. Spec. Educ. Technol. 36, 77–83. doi: 10.1177/0162643421994266

Crossref Full Text | Google Scholar

El-Ariss, B., Zaneldin, E., and Ahmed, W. (2021). Using videos in blended e-learning for a structural steel design course. Educ. Sci. 11:290. doi: 10.3390/educsci11060290

Crossref Full Text | Google Scholar

Guo, P. J., Kim, J., and Rubin, R.. 2014. How video production affects student engagement: an empirical study of MOOC videos. Proceedings of the First ACM conference on learning@ scale conference, 41–50.

Google Scholar

Izani, M., Assad, A., Kaleel, A., Wong, L., Abdulla, D., and Hamdan, A. (2025). “Evaluating AI-generated video quality: a novel assessment model” in Intelligent computing systems. eds. A. Safi, A. Martin-Gonzalez, C. Brito-Loeza, and V. Castañeda-Zeman (Cham: Springer Nature Switzerland), 54–66.

Google Scholar

Haddaway, N. R., Collins, A. M., Coughlin, D., and Kirk, S. (2015). The role of Google Scholar in evidence reviews and its applicability to grey literature searching. PloS One. 10, e0138237. doi: 10.1371/journal.pone.0138237

Crossref Full Text | Google Scholar

Lan, H. T. Q., Long, N. T., and Hanh, N.Van 2020 Validation of depression, anxiety and stress scales (Dass-21): immediate psychological responses of students in the e-learning environment Int. J. High. Educ. 9, 125–133 doi: 10.5430/ijhe.v9n5p125

Crossref Full Text | Google Scholar

Lazarevic, B. (2011). Examining the role of the introductory video in the development of teaching presence in online instruction [University of Nebraska-Lincoln]. Available online at: https://digitalcommons.unl.edu/teachlearnstudent/10/ (Accessed June 15, 2025).

Google Scholar

Leiker, D., Gyllen, A. R., Eldesouky, I., and Cukurova, M. (2023). “Generative AI for learning: investigating the potential of learning videos with synthetic virtual instructors” in Artificial intelligence in education. Posters and late breaking results, workshops and tutorials, industry and innovation tracks, practitioners, doctoral consortium and blue sky. eds. N. Wang, G. Rebolledo-Mendez, V. Dimitrova, N. Matsuda, and O. C. Santos (Cham: Springer Nature Switzerland), 523–529.

Google Scholar

Long, N. T., and Hanh, N.Van 2020 A structural equation model of blended learning culture in the classroom Int. J. High. Educ. 9 99–115 doi: 10.5430/ijhe.v9n4p99

Crossref Full Text | Google Scholar

Mir, K., Iqbal, M. Z., and Shams, J. A. (2021). Investigation of students’ satisfaction about H5P interactive video on moodle for online learning. Int. J. Distance Educ. E-Learn. 7, 71–82. doi: 10.36261/ijdeel.v7i1.2228

Crossref Full Text | Google Scholar

Miranda, S., and Vegliante, R. (2025). Leveraging AI-generated virtual speakers to enhance multilingual e-learning experiences. Information 16:132. doi: 10.3390/info16020132

Crossref Full Text | Google Scholar

Mo, C.-Y., Wang, C., Dai, J., and Jin, P. (2022). Video playback speed influence on learning effect from the perspective of personalized adaptive learning: a study based on cognitive load theory. Front. Psychol. 13:839982. doi: 10.3389/fpsyg.2022.839982,

PubMed Abstract | Crossref Full Text | Google Scholar

Mogavi, R. H., Wang, D., Tu, J., Hadan, H., Sgandurra, S. A., Hui, P., et al. (2024). Sora OpenAI’s prelude: social media perspectives on Sora OpenAI and the future of AI video generation. ArXiv Preprint ArXiv:2403.14665. doi: 10.48550/arXiv.2403.14665

Crossref Full Text | Google Scholar

Mohamed, A. A., and Lucke-Wold, B. (2024). Text-to-video generative artificial intelligence: sora in neurosurgery. Neurosurg. Rev. 47:272. doi: 10.1007/s10143-024-02514-w,

PubMed Abstract | Crossref Full Text | Google Scholar

Netland, T., von Dzengelevski, O., Tesch, K., and Kwasnitschka, D. (2025). Comparing human-made and AI-generated teaching videos: an experimental study on learning effects. Comput. Educ. 224:105164. doi: 10.1016/j.compedu.2024.105164

Crossref Full Text | Google Scholar

Pellas, N. (2023). The influence of sociodemographic factors on students’ attitudes toward AI-generated video content creation. Smart Learn. Environ. 10:57. doi: 10.1186/s40561-023-00276-4

Crossref Full Text | Google Scholar

Sanal Kumar, T. S., and Thandeeswaran, R. (2025). An improved adaptive personalization model for instructional video-based e-learning environments. J. Comput. Educ. 12, 267–313. doi: 10.1007/s40692-023-00310-x,

PubMed Abstract | Crossref Full Text | Google Scholar

Shu, J., Kou, H., Zhang, J., Zhang, T., Zhang, H., Xu, T., et al. (2025). AI-generated instructional videos: a systematic review of learning impacts, applications, and research perspectives. 2025 7th international conference on computer science and technologies in education (CSTE), 877–881.

Google Scholar

Struger, P., Brünner, B., and Ebner, M. 2025. Synthetic educators: analyzing AI-driven avatars in digital learning environments. International conference on human–computer interaction, 187–199. Doi: doi: 10.1007/978-3-031-93567-1_13

Crossref Full Text | Google Scholar

Temsah, M.-H., Nazer, R., Altamimi, I., Aldekhyyel, R., Jamal, A., Almansour, M., et al. (2025). Openai’s Sora and Google’s Veo 2 in action: a narrative review of artificial intelligence-driven video generation models transforming healthcare. Cureus 17:e77593. doi: 10.7759/cureus.77593,

PubMed Abstract | Crossref Full Text | Google Scholar

Tonbuloğlu, B., and Tonbuloğlu, İ. (2023). Trends and patterns in blended learning research (1965–2022). Educ. Inf. Technol. 28, 13987–14018. doi: 10.1007/s10639-023-11754-0,

PubMed Abstract | Crossref Full Text | Google Scholar

Tuyet, N. T., Long, N. T., and Van Hanh, N. (2020). The effect of positive learning culture in students’ blended learning process. J. E-Learn. Knowl. Soc. 16, 68–75. doi: 10.20368/1971-8829/1135218

Crossref Full Text | Google Scholar

Vallis, C., Wilson, S., Gozman, D., and Buchanan, J. (2024). Student perceptions of AI-generated avatars in teaching business ethics: we might not be impressed. Postdigit. Sci. Educ. 6, 537–555. doi: 10.1007/s42438-023-00407-7

Crossref Full Text | Google Scholar

Waisberg, E., Ong, J., Kumar, R., Masalkhi, M., and Lee, A. G. (2024). Openai’s sora in medical education: artificial videos in the classroom of the future. Can. Med. Educ. J. doi: 10.36834/cmej.79065

Crossref Full Text | Google Scholar

Xie, T., Zheng, Q., Zhang, W., and Qu, H. (2017). Modeling and predicting the active video-viewing time in a large-scale e-learning system. IEEE Access 5, 11490–11504. doi: 10.1109/ACCESS.2017.2717858

Crossref Full Text | Google Scholar

Xu, T., Liu, Y., Jin, Y., Qu, Y., Bai, J., Zhang, W., et al. (2025). From recorded to AI-generated instructional videos: a comparison of learning performance and experience. Br. J. Educ. Technol. 56, 1463–1487. doi: 10.1111/bjet.13530

Crossref Full Text | Google Scholar

Yu, T., Yang, W., Xu, J., and Pan, Y. (2024). Barriers to industry adoption of AI video generation tools: a study based on the perspectives of video production professionals in China. Appl. Sci. 14:5770. doi: 10.3390/app14135770

Crossref Full Text | Google Scholar

Zhang, D., Zhou, L., Briggs, R. O., and Nunamaker, J. F. (2006). Instructional video in e-learning: assessing the impact of interactive video on learning effectiveness. Inf. Manag. 43, 15–27. doi: 10.1016/j.im.2005.01.004

Crossref Full Text | Google Scholar

Zhou, K. Z., Choudhry, A., Gumusel, E., and Sanfilippo, M. R. (2024). ‘Sora is incredible and scary’: public perceptions and governance challenges of text-to-video generative AI models. Inf. Res. Int. Electron. J. 30, 508–522. doi: 10.47989/ir30iConf47290

Crossref Full Text | Google Scholar

Keywords: video-based learning, artificial intelligence, AI-generated instructional videos, AI-generated videos, higher education

Citation: Hai TT, Mai DTT and Hanh NV (2026) A rapid review of using AI-generated instructional videos in higher education. Front. Comput. Sci. 7:1721093. doi: 10.3389/fcomp.2025.1721093

Received: 09 October 2025; Revised: 11 November 2025; Accepted: 05 December 2025;
Published: 06 January 2026.

Edited by:

Songxin Tan, South Dakota State University, United States

Reviewed by:

Potsirin Limpinan, Rajabhat Maha Sarakham University, Thailand
Alessio Faccia, University of Birmingham Dubai, United Arab Emirates

Copyright © 2026 Hai, Mai and Hanh. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Nguyen Van Hanh, SGFuaC5uZ3V5ZW52YW5AaHVzdC5lZHUudm4=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.