- 1Department of Endodontics, RAK College of Dental Sciences, RAK Medical and Health Sciences University, Ras Al-Khaimah, United Arab Emirates
- 2Department of Endodontics, Faculty of Dentistry, Assiut University, Assiut, Egypt
- 3Department of Clinical Sciences, College of Dentistry, Ajman University, Ajman, United Arab Emirates
- 4Centre of Medical and Bio-allied Health Sciences Research, Ajman, United Arab Emirates
- 5Postgraduate Program in Health Sciences, University of Taubate, Taubaté, Brazil
- 6Department of Community Medicine, RAK College of Medicine Sciences, RAK Medical and Health Sciences University, Ras Al Khaimah, United Arab Emirates
- 7Department of Clinical Sciences, RAK College of Dental Sciences, RAK Medical and Health Sciences University, Ras Al-Khaimah, United Arab Emirates
- 8Department of Oral Radiology, RAK College of Dental Sciences, RAK Medical and Health Sciences University, Ras Al-Khaimah, United Arab Emirates
- 9Department of Conservative Dentistry and Endodontics, Saveetha Dental College and Hospitals, Saveetha Institute of Medical and Technical Sciences, Saveetha University, Chennai, Tamil Nadu, India
A profound understanding of dental students' use of artificial intelligence (AI) applications and their perceptions is essential for promoting responsible adoption and guiding integration into dental education.
Aim: This study evaluated dental students' perceptions, usage patterns, and trust in ChatGPT, Snapchat's My AI, and the Metaverse as educational tools.
Materials and methods: A cross-sectional online survey was administered to undergraduate dental students at RAK College of Dental Sciences, United Arab Emirates, between February and May 2024. The questionnaire comprised 29 questions organized into five sections. The questions were formulated as multiple-choice questions. Descriptive statistics were calculated. Pearson's chi-square tests with post-hoc adjusted standardized residuals (ASRs) and Bonferroni corrections examined categorical associations. Ordinal logistic regression assessed predictors of AI awareness, and paired-samples t-tests with Cohen's d compared trust in ChatGPT vs. Snapchat AI.
Results: The response rate was 57%, of which 70% were females. Self-rated awareness of AI increased across academic years, with fifth-year students rating their awareness score as 4 ± 0.9 out of 5, compared to 3 ± 1.4 in the first year. Ordinal regression showed no significant effects of year, gender, or their interaction (p > 0.19). ChatGPT was the most used tool (81.5%), followed by grammar correction tools (75%) and Snapchat AI (74.4%), while Metaverse use was limited (28.6%). Chi-square analyses confirmed significantly greater use of ChatGPT, Grammarly, and Snapchat AI compared with Metaverse (p < 0.001). Fourth-year students most often used AI for academic or clinical purposes (41.7%). Educational potential was endorsed by 78% of students, while privacy and data security were the predominant concerns (78%). Compared with Snapchat AI, ChatGPT was significantly more often used for education, preferred for quick responses, and more frequently associated with positive beliefs about future learning. Paired-samples t-tests demonstrated consistently higher trust in ChatGPT across all academic years, with moderate-to-large effect sizes (Cohen's d = 0.67–0.91; p < 0.01).
Conclusions: Dental students reported widespread adoption and higher trust in ChatGPT compared to other AI tools. While recognizing its educational potential, concerns about accuracy and privacy underscore the need for integrating AI literacy and evidence-based evaluation into dental curricula.
Introduction
Artificial intelligence (AI) has rapidly expanded across healthcare and education, offering tools that enhance information retrieval, problem-solving, and decision support (1). AI is a transdisciplinary field that involves the use of computer algorithms to model intelligent behaviors with minimal human intervention. AI applications are primarily based on machine learning and utilize information retrieval, image recognition, and cognitive decision-support systems (2). In dentistry, AI has demonstrated applications in radiographic interpretation, diagnostic support, treatment planning, and simulation-based learning (3). Beyond clinical practice, AI has the potential to transform dental education by providing students with interactive platforms, immediate feedback, and supplementary learning resources that complement traditional curricula (4).
Among available AI applications, ChatGPT (OpenAI, San Francisco, USA) has gained significant attention for its ability to generate coherent, human-like responses and assist with academic tasks. Several studies have shown that dental students and educators are increasingly using ChatGPT for literature support, assignment preparation, and clarification of complex concepts (5–7). While its accessibility and versatility are attractive, concerns persist regarding the accuracy and reliability of its outputs, particularly in healthcare education, where evidence-based information is critical (8). Similarly, Snapchat's My AI chatbot, powered by the same underlying language model, has become widely available to students as a mobile application. Although designed for general use, its easy accessibility raises questions about whether students employ it for academic support and how it compares with ChatGPT in terms of trust and perceived utility (9). In addition, the Metaverse, which integrates virtual and augmented reality, has been increasingly explored in medical and dental training as an immersive tool for clinical simulation and skill acquisition (10). The application of the metaverse in education still faces challenges such as cost, accessibility, privacy, and data security (11).
Despite the potential advantages of these AI applications in dental education, these tools may lack the depth, accuracy, and contextual relevance of information provided by trusted sources such as textbooks and evidence-based educational materials (12). While prior studies have examined dental students' awareness and attitudes toward AI (13, 14), most have focused on general perceptions rather than specific applications currently available to students.
Understanding the current usage patterns of these applications among dental students, as well as the extent of their frequency of use and their level of awareness of AI tools, will greatly assist in incorporating these applications into dental curricula in the future. The present study aims to evaluate the perceptions, usage patterns, and trust of dental students in relation to ChatGPT, My AI Snapchat, and the Metaverse. Specifically, it investigates the frequency of use of these applications for academic purposes, the types of information sought, and students' subjective assessment of the accuracy and reliability of the information received. The findings may inform strategies to integrate AI tools into dental education in a manner that enhances learning while upholding academic rigor and reliability.
Methods
A comprehensive questionnaire was developed in the English language following a thorough analysis of questions used in previous studies with similar objectives (15). The formulation process adhered to specific guidelines for survey reporting (16) and was overseen by experts in the field. The study was conducted in accordance with the Declaration of Helsinki, and the proposal was approved by the Institutional Review Board (RAKCODS-REC−18−2023/24-UG). The self-administered questionnaire was hosted on Google Forms and distributed along with a cover letter clearly outlining the study's objectives and the measures taken to safeguard participant anonymity. To ensure the clarity, validity, and effectiveness of the questionnaire, it underwent rigorous evaluation by two dental professors. Participants were assured of their anonymity in the survey, and the invitation explicitly stated that there were no incentives or penalties associated with participation in the study. Before the main survey was conducted, a pilot test was carried out on a small group of students to assess the clarity, feasibility, and applicability of the questionnaire. The link to the Google Form was emailed to all dental students twice between February and May 2024.
Questionnaire structure
The Google Forms questionnaire comprised 29 questions organized into five sections. Questions are formulated as multiple-choice questions. The introduction clarifies the research title and aims of the study. A mandatory informed consent was introduced: “Do you agree to participate in this study?” with “yes” as the answer being required to move into the next section of the survey, while the answer of “no” resulted in the closure of the survey. The first section was demographic information and included two questions, which were gender and the grade of dental education. In the second section, the participants were asked six questions about their awareness and experience of using various AI tools. The third section comprised 9 questions regarding the perceived benefits and challenges of AI in dental education. The fourth section was related to the specific use and experience of ChatGPT and included six questions. The fifth section was specific to the use and experience of Snapchat's My AI feature and included six questions (Supplementary Appendix S1).
Statistical and data analysis
Data were analyzed using IBM SPSS Statistics, version 29 (IBM Corp., Armonk, NY, USA). Descriptive statistics were computed for all variables and are presented as frequencies and percentages for categorical data and as mean ± standard deviation (SD) or median with interquartile range (IQR) for continuous and ordinal data, as appropriate. The required sample size was calculated using a 95% confidence level, 5% margin of error, estimated population proportion of 50%, and a total population size of 295 students, yielding a minimum target of 167 participants.
Associations between categorical variables, such as AI tool usage status, frequency of use, and reasons for preference, and beliefs about educational potential, were examined using Pearson's chi-square tests. Contingency tables were constructed for comparisons, including (i) AI tool (ChatGPT, Snapchat AI, Metaverse, Grammarly) × usage status (ever vs. never used), (ii) AI tool × reason for preference (quick responses, trustable answers, not used), and (iii) software type (ChatGPT vs. Snapchat AI) × beliefs about enhancing dental education (strongly disagree to strongly agree). For tables yielding significant omnibus chi-square results, post-hoc analyses were performed using adjusted standardized residuals (ASRs). Bonferroni corrections were applied to control for multiple comparisons, with adjusted thresholds (e.g., ±2.73 for 2 × 4 tables and ±2.81 for 2 × 5 tables) used to determine significance.
Paired-samples t-tests were performed within each academic year to directly compare trust ratings between ChatGPT and Snapchat AI. For comparing paired trust scores between ChatGPT and Snapchat AI, both parametric and non-parametric tests were applied. Normality of within-subject differences (ChatGPT—Snapchat AI) was verified using Q–Q plots and the Shapiro–Wilk test. As the distribution deviated from normality (p < 0.05), a Wilcoxon signed-rank test was additionally performed as a sensitivity analysis to confirm the robustness of results. The paired-samples t-test was retained as the primary analysis, given the sufficiently large sample size and the robustness of the t-test to moderate deviations from normality.
Effect sizes were calculated using Cohen's d, along with 95% confidence intervals (CIs). Since five paired t-tests were conducted to compare ChatGPT and Snapchat AI trust scores across academic years (Year 1 to Year 5), each set of yearly comparisons was considered a single family of related tests. To control for type I error inflation due to multiple testing, a Bonferroni correction was applied. The adjusted significance threshold for these comparisons was set at α = 0.05/5 = 0.01. Within each family, p-values below 0.01 were considered statistically significant. To further investigate factors influencing students' trust and awareness of AI-based software tools, a series of ordinal logistic regression models using the cumulative logit link function were employed. Separate models were developed for each dependent variable: trust in ChatGPT, trust in Snapchat AI, and general AI awareness. The predictor variables included year of study, gender, and their interaction term (Year × Gender) to examine potential combined effects. Model fit was evaluated using the −2 Log Likelihood ratio test, along with Pearson and Deviance χ2 statistics to assess overall model adequacy. The Test of Parallel Lines was applied to confirm the proportional odds assumption. Nagelkerke R2 was reported as a measure of the variance explained by the model. Results included β coefficients (log-odds), Wald χ2 statistics, p-values, and Adjusted Odds Ratios (AOR = e^β) with corresponding 95% confidence intervals.
Results
Out of 295 students invited to participate in the study, 168 responded, yielding a response rate of 57%. Among the respondents, 70% were females (n = 118) and 30% were males (n = 50). The distribution of participants across academic years showed that first-year students comprised the largest group at 32% (n = 53), followed by third-year students at 20% (n = 33), second-year students at 19% (n = 32), fifth-year students at 15% (n = 26), and fourth-year students at 14% (n = 24) (Figure 1). Analysis of self-rated awareness of AI applications measured on a 5-point Likert scale showed a progressive increase across academic years, with fifth-year students reporting the highest median score (4, IQR = 3–5) compared to first-year students (3, IQR = 2–5) (Figure 2). Mann–Whitney U tests revealed no significant gender differences in awareness either within individual years or overall (U = 3,164.0, p = 0.445) (Table 1). Ordinal logistic regression was performed with AI awareness as the dependent variable and year of study, gender, and their interaction as predictors. None of the interaction terms between year and gender reached statistical significance (all p-values >0.19). For instance, awareness levels among fifth-year females (β = 1.046, 95% CI: −0.543 to 2.635, p = 0.197) and fifth-year males (β = 0.961, 95% CI: −0.624 to 2.546, p = 0.235) did not significantly differ from the reference group (third-year males). The overall model was not statistically significant, χ2(9) = 9.34, p = 0.41, and accounted for only a small proportion of the variance (Nagelkerke R2 = 0.057) (Table 2). These results suggest that year of study, gender, and their interaction did not significantly predict dental students' awareness of AI applications.
Figure 2. Self-rated awareness of artificial intelligence (AI) applications among dental students. Awareness was assessed using a 5-point Likert scale (1 = very low; 5 = very high).
Table 1. Mean ± standard deviation (SD) scores of dental students’ self-rated awareness of artificial intelligence (AI) applications on a 5-point Likert scale (1 = very low; 5 = very high).
Table 2. Ordinal logistic regression predicting AI awareness by year of study, gender, and their interaction.
With respect to AI tool usage, ChatGPT was the most frequently used application (81.5%), followed by grammar correction tools (75%) and Snapchat AI (74.4%). In contrast, only 28.6% had experience with the Metaverse. Significant differences were observed in reported usage of AI applications (χ2 = 131.9, df = 3, p < 0.0001). post-hoc residual analysis with Bonferroni correction (α = 0.00625; threshold ±2.73) revealed that ChatGPT, Grammarly, and Snapchat AI were used significantly more often than expected, whereas Metaverse was used significantly less often (Table 3). Interestingly, first-year students had the highest percentage of metaverse usage (32.1%), while fifth-year students had the lowest (15.4%) (Figure 3).
Table 3. Frequencies and percentages of dental students who reported ever using or never using four artificial intelligence (AI) applications (ChatGPT, snapchat AI, metaverse, and grammar correction tools).
Figure 3. Academic year-wise use of AI applications by dental students. Proportions of students reporting ever or never using ChatGPT, Snapchat's My AI, the Metaverse, or grammar correction tools.
When asked about the use of AI tools as resources for studies or dental practice, fourth-year students reported the highest utilization rate (41.7%), followed by second-year students (28.1%), with the lowest utilization seen among first-year students (17%). However, the chi-square test found no statistically significant association between AI usage and academic year (χ2 = 6.87, p = 0.143) (Table 4).
Table 4. Frequencies and percentages of dental students who reported using any AI tool as an information resource for academic studies or clinical dental practice, stratified by academic year.
Regarding students' perceptions of AI applications in dentistry, 78% believed that AI has potential in dental education, 60.7% identified its usefulness in treatment planning, and 48% saw potential in research and data analysis. Conversely, administrative tasks were the least selected domain, with only 19% of students recognizing their potential (Figure 4). When addressing concerns about AI integration into dentistry, the most common issue was data security and patient privacy (78%), whereas only 31% expressed concern about AI reducing the role of dental practitioners in the future (Figure 5).
Figure 4. Perceived potential applications of AI in dentistry. Participants indicated the aspects of dentistry where they believed AI has the most potential.
Comparisons between ChatGPT and Snapchat AI demonstrated notable differences. For educational use, ChatGPT was significantly more often reported as a source than Snapchat AI (χ2 = 21.7, df = 4, p < 0.001). post-hoc analysis with Bonferroni correction (α = 0.005, cutoff ±2.81) indicated that significantly fewer students reported “never” using ChatGPT for education, whereas significantly more students reported “never” using Snapchat AI. Regarding reasons for preference, a significant association was found between the main reason for preferring an AI tool and the type of software (χ2 = 21.61, df = 2, p < 0.001). Post-hoc analysis with Bonferroni correction (α = 0.0083, cutoff ±2.64) revealed that significantly more students (63.7%) reported ChatGPT as providing “quick and effortless responses” (ASR > +2.7), whereas significantly more students reported Snapchat AI as “not used as a source of information” (ASR > +2.8). No significant difference was observed for “trustable answers”.
Students' perceptions of privacy concerns did not differ significantly between ChatGPT and Snapchat AI (χ2 = 6.58, p = 0.08). However, beliefs about the potential of these tools to enhance future learning showed marked differences: 50.6% of participants agreed or strongly agreed that ChatGPT could positively impact dental education, compared with only 16.7% for Snapchat AI. A significant association was found between the type of software and belief about enhancing learning in dental education (χ2 = 50.39, p < 0.0001). post-hoc analysis with Bonferroni correction (α = 0.005, cutoff ±2.81) indicated that ChatGPT was significantly more often associated with “Agree,” while Snapchat AI was more often associated with “Neutral” and “Strongly disagree” responses (Table 5). Trust scores for ChatGPT were consistently higher across all academic years compared to Snapchat AI. Paired-samples t-tests revealed that, across all academic years, students consistently reported significantly greater trust in ChatGPT compared to Snapchat AI (all p < 0.001, except for fourth-year students, p = 0.003). The observed effect sizes ranged from moderate to large (Cohen's d = 0.67–0.91). After applying Bonferroni correction (α = 0.01 for five paired comparisons), all results for trust differences between ChatGPT and Snapchat AI remained statistically significant (Table 6).
Table 5. Frequencies and percentages of dental students’ specific use and attitudes regarding ChatGPT and snapchat AI, including frequency of use, educational applications, reasons for preference, privacy concerns, and beliefs about future impact on dental education.
Table 6. Paired samples t-test comparing mean (±SD) trust scores for ChatGPT and snapchat AI among dental students across years of study, measured on a 5-point Likert scale (1 = very low; 5 = very high).
An ordinal logistic regression was performed to assess the effects of year of study, gender, and their interaction on levels of trust in ChatGPT. The overall model was statistically significant, χ2(9) = 21.78, p = 0.010, indicating that the inclusion of predictors significantly improved model fit. Goodness-of-fit indicators were acceptable, with Pearson χ2(27) = 21.43, p = 0.77 and Deviance χ2(27) = 25.86, p = 0.53, suggesting adequate model representation. The proportional-odds assumption was met (p = 0.357). The Nagelkerke R2 was 0.13, indicating modest explanatory power. Among predictors, only the interaction between being a fourth-year and male student was statistically significant (β = −2.13, p = 0.028, AOR = 0.12, 95% CI: 0.02–0.79), suggesting that fourth-year male students were significantly less likely to report higher levels of trust in ChatGPT compared with third-year male students. No other main effects or interactions reached statistical significance (Table 7).
Table 7. Ordinal logistic regression predicting trust in ChatGPT by year of study, gender, and their interaction.
A second ordinal logistic regression model with a cumulative logit link examined students' trust in Snapchat AI as a function of year of study, gender, and their interaction. The overall model was not statistically significant, χ2(9) = 12.17, p = 0.204, indicating that the inclusion of predictors did not significantly improve model fit over the null model. Goodness-of-fit statistics were acceptable [Pearson χ2(36) = 26.95, p = 0.86; Deviance χ2(36) = 33.65, p = 0.58], and the proportional-odds assumption was met (Test of Parallel Lines: p = 0.073). The model's explanatory power was limited (Nagelkerke R2 = 0.07). Notably, two interaction terms were statistically significant: first-year female students (β = −1.41, p = 0.044, AOR = 0.24, 95% CI: 0.06–0.96) and third-year female students (β = −1.48, p = 0.045, AOR = 0.23, 95% CI: 0.05–0.97) were significantly less likely to report higher trust in Snapchat AI compared with third-year male students. All other year-by-gender interaction terms were non-significant (Table 8).
Table 8. Ordinal logistic regression predicting trust in snapchat AI by year of study, gender, and their interaction.
Discussion
Rapid advancements in artificial intelligence (AI) and its adoption for teaching and educational purposes could mark a new era in academia (17). Currently, the most popular AI tool among students worldwide is ChatGPT, which is considered the most rapidly expanding consumer application in history (18). In parallel, the Snapchat AI application has become available on mobile phones, particularly among the age group of undergraduate students. Metaverse has also gained significant attention in the education sector in the last two years (11). To our knowledge, this is the first survey-based study conducted among undergraduate dental students in the UAE to assess awareness, perceptions, and trust in these popular AI tools and applications.
In our study, awareness of AI applications increased across academic years, with final-year students reporting the highest confidence in their knowledge. However, ordinal logistic regression revealed that neither year of study nor gender significantly predicted awareness, suggesting that exposure alone may not determine familiarity. Interestingly, first-year students reported the highest percentage of utilization of ChatGPT (92.5%) and the highest rate of awareness of the Metaverse (32.1%), which may reflect curiosity or early adoption trends among new cohorts. Previous research indicates that students might overestimate their level of awareness about AI, and this should be taken into consideration in the future dental curriculum (19).
When it comes to using AI tools specifically as resources for studies or dental practice, fourth-year students had the highest utilization rate (41.7%). This trend may be explained by the curriculum structure at our dental college, where the fourth year primarily focuses on theoretical coursework, while the fifth year emphasizes practical and clinical training. In this study, the majority of students (78%) believed that AI has high potential in dental education. This comes in agreement with a recent study conducted in Jordan by Ajlouni et al., which showed that a majority of students (73%) agreed on the potential of ChatGPT in facilitating the learning process (20). These findings are similar to those of a study conducted in Saudi Arabia, which reported that 72% of the participants believed that AI should be incorporated into postgraduate and undergraduate dental education (19). Of the respondents, 60.7% believed that AI could be very helpful in treatment planning. Most of the results from previous studies are more or less similar to the results of the current survey (15, 21).
The primary concern among students regarding AI integration into dentistry was related to data security and patient privacy. Notably, only 31% expressed apprehension that AI might diminish the role of dental practitioners in the future, aligning with the findings of a previous survey conducted across nine Turkish dental schools (15). The findings of similar studies conducted in Saudi Arabia, but with higher concern levels reported, with almost 50%–64% agreeing that AI could replace dentists in the future (19, 22). The conclusion in the literature is that the dental field is distinct from other professions in which the human element can be replaced by algorithms. In dentistry, human traits such as accumulated experience, empathy with patients, and the trust-based relationships between patients and dentists cannot be entirely substituted by AI-operated robots (23). Different age groups of university students may have varying needs, preferences, and levels of familiarity with technological advancements.
Comparisons between ChatGPT and Snapchat AI revealed significant differences in use and perception. post-hoc analyses of adjusted standardized residuals indicated that ChatGPT was more often associated with “quick and effortless responses,” while Snapchat AI was more commonly reported as “not used as a source of information.” Trust was consistently higher for ChatGPT across all years. Students also expressed greater optimism about ChatGPT's role in enhancing dental education, whereas Snapchat AI was more frequently linked to neutral or negative views. These findings highlight that while both tools are accessible, students regard ChatGPT as more credible and academically relevant. In a previous multinational cross-sectional study, the majority of students intended to use ChatGPT for assignment support and anticipated that their peers would endorse its usage, implying a potential shift towards ChatGPT use becoming a standard practice among university students (24). In an exploratory study investigating the performance of ChatGPT on a wide range of dental assessments and discussing the implications for undergraduate dental education, ChatGPT provided accurate responses to the majority of knowledge-based assessments. It was found that ChatGPT has the potential to revolutionize virtual learning. The authors recommend that dental educators adapt their teaching and assessments with this tool to the benefit of the learners (25). In another study, ChatGPT was found to increase students' motivation to learn (26).
Despite the advancements of ChatGPT, it is critical to recognize its limitations (27). While it is a powerful tool for synthesizing information and enhancing learning experiences, it operates based on patterns learned from training data and lacks the ability to critically evaluate or verify information (28). This absence of inherent reliability, particularly in clinical, medical, or scientific contexts, poses a potential risk to dental students who might use it as a primary source of information. Non-evidence-based or inaccurate content can inadvertently influence their learning and decision-making processes, emphasizing the need for cautious and supplementary use alongside validated resources (29, 30). For dental educators, this underscores the importance of teaching students to critically evaluate AI-generated content and triangulate it against validated sources.
It is important to acknowledge certain limitations of this study. First, the survey was conducted at a single dental school in the UAE, which restricts the generalizability of the findings. The data may not represent the perspectives of all dental students across the country, particularly given the differences in institutional curricula, learning environments, and exposure to digital technologies. Second, the participant pool consisted predominantly of female students (70%), which introduces a potential gender bias; the results may vary if a balanced representation of genders is achieved. Although this imbalance is a limitation, it is noteworthy that similar demographic trends have been reported in previous studies, including a study conducted in the UAE, in which female students represented 68% (31), as well as in Spanish dental schools (71%) (32), India (79.5%) (33), and Saudi Arabia; this predominance likely reflects the global demographic shift toward greater female enrollment in dental education (34). Finally, the cross-sectional design limits causal inference.
Future research should aim to address these limitations by recruiting larger and more diverse samples, encompassing all dental schools in the UAE, and ensuring balanced representation of genders to explore possible differences in preferences and attitudes. Studies incorporating sufficient covariates will enable multivariable adjustment and more robust analyses. Additionally, subsequent studies should look deeper into assessing the current knowledge and attitudes towards AI tools among dental students. This will provide a richer understanding of AI's potential impact and applications of AI in dental education and practice.
Conclusion
The rise of AI-driven technologies, such as ChatGPT, Snapchat's My AI, and virtual platforms like the Metaverse, heralds a new era in education with the potential to redefine traditional learning methods. These tools offer diverse resources and enable interactive learning experiences for the students. In the context of dental education, integrating topics on the application of AI tools into the curriculum is becoming increasingly essential. Emphasis should also be placed on fostering critical evaluation skills, ensuring that students can validate AI-provided information and align it with the principles of evidence-based dentistry.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
The studies involving humans were approved by Research and Ethics Committee, RAKMHSU, UAE. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributions
ME: Investigation, Software, Validation, Formal analysis, Methodology, Writing – original draft, Data curation. KB: Writing – review & editing, Methodology, Investigation, Formal analysis. MP: Methodology, Writing – original draft, Investigation. ME: Writing – original draft, Investigation, Methodology. NC: Methodology, Validation, Writing – review & editing, Formal analysis. MK: Validation, Methodology, Formal analysis, Supervision, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research and/or publication of this article. Ajman University supports the publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fdmed.2025.1673536/full#supplementary-material
References
1. King MR. The future of AI in medicine: a perspective from a Chatbot. Ann Biomed Eng. (2023) 51(2):291–5. doi: 10.1007/s10439-022-03121-w
2. Howard J. Artificial intelligence: implications for the future of work. Am J Ind Med. (2019) 62(11):917–26. doi: 10.1002/ajim.23037
3. Jha Kukreja B, Kukreja P. Integration of artificial intelligence in dentistry: a systematic review of educational and clinical implications. Cureus. (2025) 17(2):e79350. doi: 10.7759/cureus.79350
4. Fang Q, Reynaldi R, Araminta AS, Kamal I, Saini P, Afshari FS, et al. Artificial intelligence (AI)-driven dental education: exploring the role of Chatbots in a clinical learning environment. J Prosthet Dent. (2024) 134:1296–303.38644064
5. Massey PA, Montgomery C, Zhang AS. Comparison of ChatGPT–3.5, ChatGPT-4, and orthopaedic resident performance on orthopaedic assessment examinations. J Am Acad Orthop Surg. (2023) 31(23):1173–9. doi: 10.5435/JAAOS-D-23-00396
6. Munaye YY, Admass W, Belayneh Y, Molla A, Asmare M. ChatGPT in education: a systematic review on opportunities, challenges, and future directions. Algorithms. (2025) 18(6):352. doi: 10.3390/a18060352
7. Foo LG, Rahmani H, Liu J. AI-generated content (AIGC) for various data modalities: a survey. ACM Comput Surv. (2025) 57(9):1–66. doi: 10.1145/3728633
8. Ali K, Barhom N, Tamimi F, Duggal M. ChatGPT—a double-edged sword for healthcare education? Implications for assessments of dental students. Eur J Dent Educ. (2024) 28(1):206–11. doi: 10.1111/eje.12937
9. Antala AV, Kariya PB. Social media as a tool for oral health education in early childhood and adolescents: opportunities and challenges. Arch Med Health Sci. (2025) 10:4103. doi: 10.4103/amhs.amhs_152_24
10. Thurzo A, Strunga M, Urban R, Surovková J, Afrashtehfar KI. Impact of artificial intelligence on dental education: a review and guide for curriculum update. Educ Sci. (2023) 13(2):150. doi: 10.3390/educsci13020150
11. Park HJ, Kim EJ, Kim JY. Exploring large language models and the metaverse for urologic applications: potential, challenges, and the path forward. Int Neurourol J. (2024) 28(Suppl 2):S65–73. doi: 10.5213/inj.2448402.201
12. Uribe SE, Maldupa I, Schwendicke F. Integrating generative AI in dental education: a scoping review of current practices and recommendations. Eur J Dent Educ. (2025) 29(2):341–55. doi: 10.1111/eje.13074
13. Hegde S, Nanayakkara S, Jordan A, Jeha O, Patel U, Luu V, et al. Attitudes and perceptions of Australian dentists and dental students towards applications of artificial intelligence in dentistry: a survey. Eur J Dent Educ. (2025) 29(1):9–18. doi: 10.1111/eje.13042
14. Elchaghaby M, Wahby R. Knowledge, attitudes, and perceptions of a group of Egyptian dental students toward artificial intelligence: a cross-sectional study. BMC Oral Health. (2025) 25(1):11. doi: 10.1186/s12903-024-05282-7
15. Yüzbaşıoğlu E. Attitudes and perceptions of dental students towards artificial intelligence. J Dent Educ. (2021) 85(1):60–8. doi: 10.1002/jdd.12385
16. Bennett C, Khangura S, Brehaut JC, Graham ID, Moher D, Potter BK, et al. Reporting guidelines for survey research: an analysis of published guidance and reporting practices. PLoS Med. (2011) 8(8):e1001069. doi: 10.1371/journal.pmed.1001069
17. Crompton H, Burke D. Artificial intelligence in higher education: the state of the field. Int J Educ Technol High Educ. (2023) 20(1):22. doi: 10.1186/s41239-023-00392-8
18. Tsang R. Practical applications of ChatGPT in undergraduate medical education. J Med Educ Curric Dev. (2023) 10:23821205231178449. doi: 10.1177/23821205231178449
19. Aboalshamat KT. Perception and utilization of artificial intelligence (AI) among dental professionals in Saudi Arabia. Open Dent J. (2022) 16(1):e187421062208110. doi: 10.2174/18742106-v16-e2208110
20. Ajlouni AO, Abd-Alkareem Wahba F, Salem Almahaireh A. Students’ attitudes towards using ChatGPT as a learning tool: the case of the university of Jordan. Int J Interact Mob Technol. (2023) 17(18):99–117. doi: 10.3991/ijim.v17i18.41753
21. Ezzeldin N, Salama AA, Shehab KA. Knowledge, perception, and attitude of Egyptian dental students toward the role of robotics and artificial intelligence in dental practices-a cross-sectional study. BMC Oral Health. (2025) 25(1):747. doi: 10.1186/s12903-025-06077-0
22. ALruwail BF, Alshalan AM, Thirunavukkarasu A, Alibrahim A, Alenezi AM, Aldhuwayhi TZA. Evaluation of health science Students’ knowledge, attitudes, and practices toward artificial intelligence in northern Saudi Arabia: implications for curriculum refinement and healthcare delivery. J Multidiscip Healthc. (2025) 18:623–35. doi: 10.2147/JMDH.S499902
23. Inkster B, Sarda S, Subramanian V. An empathy-driven, conversational artificial intelligence agent (wysa) for digital mental well-being: real-world data evaluation mixed-methods study. JMIR Mhealth Uhealth. (2018) 6(11):e12106. doi: 10.2196/12106
24. Tortella F, Palese A, Turolla A, Castellini G, Pillastrini P, Landuzzi MG, et al. Knowledge and use, perceptions of benefits and limitations of artificial intelligence chatbots among Italian physiotherapy students: a cross-sectional national study. BMC Med Educ. (2025) 25(1):572. doi: 10.1186/s12909-025-07176-w
25. Aura-Tormos JI, Llacer-Martinez M, Torres-Osca I. Educational applications of ChatGPT in university-based dental education. A systematic review. Eur J Dent Educ. (2025):1–17. doi: 10.1111/eje.70011
26. Rasul T, Nair S, Kalendra D, Robin M, de Oliveira Santini F, Ladeira WJ, et al. The role of ChatGPT in higher education: benefits, challenges, and future research directions. J Appl Lear Teach. (2023) 6(1):41–56. doi: 10.37074/jalt.2023.6.1.29
27. Lin S.-M., Chung H.-H., Chung F.-L. and Lan Y.-J.: Concerns about using ChatGPT in education. In: International Conference on Innovative Technologies and Learning; Springer (2023). p. 37–49. doi: 10.1007/978-3-031-40113-8_4
28. Chan CKY. A comprehensive AI policy education framework for university teaching and learning. Int J Educ Technol High Educ. (2023) 20(1):38. doi: 10.1186/s41239-023-00408-3
29. Cooper G. Examining science education in chatgpt: an exploratory study of generative artificial intelligence. J Sci Educ Technol. (2023) 32(3):444–52. doi: 10.1007/s10956-023-10039-y
30. Wazan AS, Taj I, Shoufan A, Laborde R, Venant R. How to design and deliver courses for higher education in the AI era: insights from exam data analysis. arXiv [Preprint]. arXiv: 2308.02441v1 (2023). doi: 10.48550/arXiv.2308.02441
31. Elsayed MA, Islam MS, Saleh DR, Alnahdi AM, Padmanabhan V. Endodontic procedural errors and associated factors among undergraduate dental students: a cross-sectional study. J Contemp Dent Pract. (2023) 24(12):998–1007. doi: 10.5005/jp-journals-10024-3614
32. Martín-Jiménez M, Martín-Biedma B, López-López J, Alonso-Ezpeleta O, Velasco-Ortega E, Jiménez-Sánchez M, et al. Dental students’ knowledge regarding the indications for antibiotics in the management of endodontic infections. Int Endod J. (2018) 51(1):118–27. doi: 10.1111/iej.12778
33. Doshi A, Asawa K, Bhat N, Tak M, Dutta P, Bansal TK, et al. Knowledge and practices of Indian dental students regarding the prescription of antibiotics and analgesics. Clujul Med. (2017) 90(4):431. doi: 10.15386/cjmed-768
Keywords: ChatGPT, artificial intelligence, My AI Snapchat, Metaverse, dental students, dental education
Citation: Elsayed MA, Baroudi K, Patni MA, Elwakil MM, Chaitanya NCSK and Karobari MI (2026) Are ChatGPT, My AI Snapchat, and Metaverse used by dental students as reliable sources of dental education? Front. Dent. Med. 6:1673536. doi: 10.3389/fdmed.2025.1673536
Received: 26 July 2025; Revised: 16 October 2025;
Accepted: 25 November 2025;
Published: 6 January 2026.
Edited by:
Francisco Nociti, ADA Science & Research Institute, American Dental Association, United StatesReviewed by:
Ravindra Rapaka, University of Connecticut, United StatesRicha Kaushik, Himachal Pradesh University, India
Copyright: © 2026 Elsayed, Baroudi, Patni, Elwakil, Chaitanya and Karobari. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Kusai Baroudi, ZF9rdXNhaUB5YWhvby5jby51aw==
Mahmoud Mohamed Elwakil7