ORIGINAL RESEARCH article

Front. Commun., 26 May 2025

Sec. Culture and Communication

Volume 10 - 2025 | https://doi.org/10.3389/fcomm.2025.1479813

From assistance to integrity: exploring the role of AI in academic communication across Russian and Turkish campuses

  • 1Department of English Philology, Kuban State University, Krasnodar, Russia
  • 2Scientific and Educational Center, Person in Communication, Pyatigorsk State University, Pyatigorsk, Russia
  • 3Independent Researcher, Gaziantep, Türkiye

Introduction: This paper investigates how 33 Turkish and 38 Russian university students perceive and experience artificial intelligence (AI) in scholarly communication. This study investigates students’ perspectives, experiences, and concerns regarding AI use in educational settings. Three main areas of inquiry are addressed: general views and experiences with AI technology, effects on academic communication and teamwork, and evaluation of AI-generated work in terms of academic integrity and plagiarism.

Methods: The study participants were 71 university students, consisting of 33 Turkish students and 38 Russian students. Data was collected through open-ended questionnaires. Qualitative material was examined using six-stage theme analysis system.

Results: Our findings reveal both shared and divergent perspectives among Turkish and Russian students regarding AI’s role in their educational environment. Students from both countries recognize AI’s significant potential to streamline academic tasks and enhance access to information. However, they also voice apprehensions over its influence on critical thinking abilities and academic honesty. Turkish students exhibited a predominantly favorable outlook toward the collaborative capabilities of AI, but Russian students placed greater emphasis on apprehensions over privacy and data security.

Discussion: The study highlights the complex interplay between the benefits and challenges of AI in educational environments. Students in both countries encounter ethical issues, namely plagiarism and the authenticity of AI-generated content. Our research highlights the importance of clearly defined institutional regulations and educational initiatives to offer guidelines for the use of AI in academia. This comparative study offers a fascinating analysis of the cultural factors influencing AI applications in higher education. It adds to the ongoing worldwide conversation about how technology impacts future educational plans.

Introduction

Artificial intelligence (AI) has transformed both educational practices and communication patterns in higher education (Huang and Tan, 2023; Nikolopoulou, 2024; Thomas, 2024). Institutions are using everything from personalized learning to automated assessment systems to boost student success and institutional efficiency (Zawacki-Richter et al., 2019). However, this rapid transition raises important questions about the nature of intellectual discourse.

In academic and educational contexts, academic communication is defined as the formal, orderly means of exchanging knowledge and ideas among academics, researchers, and students. It includes scholarly papers, research reports, conference presentations, and peer reviews—written as well as oral modes of communication. The main goals of academic communication are to make sure that it increases knowledge by sharing, reviewing, and preserving scholarly work so that people in and outside of academia can access it (American Library Association, 2006; Wolfram, 2019).

New studies have shown that AI can make college more interactive and collaborative (Aldosari, 2020; Son and Jin, 2024). However, it may also bring some problems, like privacy and moral issues (Devi et al., 2023; Foltynek et al., 2023). There is limited research on the impact of AI technologies in academic communication, especially in countries such as Turkey and Russia, and these countries vary in terms of technological adoption and social norms.

This article aims to bridge this gap by looking at opinions and experiences about AI among Turkish and Russian university students. Examining how Turkish and Russian students interpret and interact with AI technologies provides insights into how cultural factors influence AI implementation in higher education.

The research questions of the study are:

1. Academic settings of university students in Turkey and Russia: AI tools;

a. purposes and frequency of use;

b. perceived benefits and limitations to investigate comparison.

2. Academic communication aided by AI in a cultural setting:

a. effects on student–student contact;

b. helps with group projects and cooperative learning;

c. offers chances for analysis of communication obstacles.

3. Students in two distinct cultural settings:

a. ethical issues for the academic usage of AI content;

b. their attitude to academic integrity and plagiarism;

c. institutional regulating expectations about the use of AI.

This study aims to improve our understanding of the cultural dimensions of the use of AI in higher education by comparing the views of Turkish and Russian students. The results can help us develop strategies that address cultural sensitivities to shape AI integration in universities.

Literature review: AI in academic communication

AI in higher education

Higher education academic contexts are increasingly incorporating AI technologies to enhance the student experience and optimize teaching and learning processes. Students are using chatbots to access services and assist in study (Chen et al., 2023; Okonkwo and Ade-Ibijola, 2021). Based on Iffath Unnisa Begum (2024), AI has grown from simple computer programs to include embedded computing systems, web-based intelligent systems, and humanoid robots that can do different tasks. AI applications in higher education enable flexible learning opportunities that are accessible anytime and anywhere (Ahmad et al., 2021; Owoc et al., 2021).

Cox (2021) argues that AI is fundamentally transforming teaching and learning in higher education, with the potential to reshape global educational systems. Tomar and Verma (2021) state that AI technologies are integral to higher education’s teaching, learning, and research activities. Zawacki-Richter et al. (2019) carefully evaluate these technologies, as they offer both positive and negative elements. Research has demonstrated that AI applications in higher education are broad, encompassing a wide range of applications, such as intelligent tutoring systems and personalized learning.

Also, research highlights the beneficial impact of AI applications on student achievement and the development of immersive educational settings (Aldosari, 2020; Jaboob et al., 2024). According to Rui and Badarch (2022), the main reason for looking into AI applications in education is to help improve teaching in higher education institutions. AI technologies are also becoming well-known in academia since they offer creative ways for research, instruction, and service (Barros et al., 2023). ChatGPT, emerging as a student-driven innovation, has the potential to enhance the higher education learning environment (Dai et al., 2023). In recent years, there has been a growing emphasis on promoting AI literacy among students, which has garnered significant interest from governments and society (Yang and Xu, 2024).

Higher education institutions are using smart classrooms and AI-enhanced learning models to progressively include AI (Jia and Tu, 2024), thereby improving the academic experience. This includes enhancing library services (Okunlaya et al., 2022) and automating administrative tasks (Hannan and Liu, 2023; Villegas-Ch et al., 2021). Top institutions are creating AI labs and implementing AI into their courses even as they assess how it may affect academic integrity and student performance (Ali et al., 2024; Sullivan et al., 2023).

From English preparation courses (Zhang and Qi, 2024) to art and design education (Watanabe, 2024; Zhao and Xue, 2024), the transformation of teaching methodologies uses big data and adaptive learning (Khan et al., 2022). According to research on AI use in engineering classes and university courses, it helps students interact with each other and learn in a more personalized way (Abdelmagid et al., 2024). Studies also reveal how AI raises student self-regulation in language acquisition and motivation (Li B. et al., 2024; Tang, 2024).

Impact of AI on academic communication

Academic communication is increasingly integrating AI, which presents both opportunities and challenges. AI tools such as ChatGPT demonstrate capabilities in paraphrasing, summarizing, and improving writing efficiency (Yan, 2023). These tools have the potential to transform scholarly communication by aiding content creation, authorship, and academic integrity (Dergaa et al., 2023; Lund and Naheem, 2024).

AI technologies, including natural language processing and big language models, have ushered in a transformative era in scholarly publishing (Lund et al., 2023; Petroșanu et al., 2023). Students believe that AI systems will increase the quantity and quality of teaching communication because they can ask more questions without feeling self-conscious or like they are interrupting the instructor. AI also allows instructors to save time by answering simple, repetitive questions and thus focus on more meaningful communication with students (Seo et al., 2021).

As AI is used more and more in research and scholarly communication, academic libraries face both opportunities and problems (Cox, 2023; Dwivedi et al., 2021). From changes in library subscriptions to the open access movement to the shift to digital publishing, AI clearly influences academic communication (Shrivastava and Mahajan, 2021). Academic publishing and research settings are undergoing major changes, thanks in great part to AI-powered technology. This is why knowledge of their consequences inside the academic community is becoming more and more important (Payini et al., 2024). Moreover, ethical use of AI in scientific content creation and education is becoming more popular, which emphasizes the need for rules and issues in using AI technology (Foltynek et al., 2023).

Although AI enhances language, vocabulary, and style (Kacena et al., 2024), ethical concerns still exist in academics, despite its advancements. Passmore and Tee (2024) analyzed AI’s opportunities in knowledge synthesis, content creation, and coaching, highlighting both its benefits and disadvantages relative to traditional methods. AI enhances student–student and student-faculty contacts in education by means of more interactive and communicative settings (Thomas, 2024). AI-powered solutions aid in permitting cooperative learning by means of real-time language translation and summarizing (Porter and Grippa, 2020; Son and Jin, 2024). Also, AI supports peer review processes by raising the quality and consistency of student work and providing insightful comments (Darvishi et al., 2022). Still, worries about AI lowering in-person interactions and maybe negative effects on social settings abound (Hohenstein et al., 2023).

Ethical considerations of AI in academia and academic integrity

With the development of AI technologies, issues of academic integrity and plagiarism have become increasingly important. AI tools such as text generators and literature review automators have facilitated academic writing but have also facilitated unethical practices such as plagiarism and cheating (Goel and Nelson, 2024; Yusuf et al., 2024). The ease with which students and researchers can create content using AI tools raises concerns about originality and citation (Cotton et al., 2024; Rabbianty et al., 2023). Generative AI poses significant threats to academic integrity by enabling practices such as plagiarism, cheating, fabrication, and unfair advantage (Song, 2024). The combination of AI and academic writing raises concerns about the reliability of research results, since fake scientific papers written by chatbots that do not do enough research can lead to issues of academic dishonesty and plagiarism (Zaitsu and Jin, 2023).

Higher education institutions are facing challenges in dealing with the widespread use of AI among students. This has led to an increase in academic misconduct due to the absence of official regulations and a lack of AI knowledge among students and instructors (Song, 2024). Academic institutions are required to revise their ethical guidelines and develop new methods for detecting plagiarism (Geethalakshmi, 2018). Additionally, it is advisable to promote knowledge about AI-supported writing and offer instruction on the ethical and appropriate utilization of these technologies (Miao et al., 2024).

It is recommended that institutions use suitable strategies to tackle plagiarism concerns and adapt to the changing environment influenced by AI technology (Bell, 2023; Khalil and Er, 2023; Perkins, 2023). Higher education institutions are being told to reevaluate and improve their academic integrity policies because AI is making it easier for more and more cases of academic misconduct to happen (Fowler, 2023). To protect academic integrity, training students and researchers on plagiarism, proper citation styles, and ethical writing is of more and more importance (Curtis and Vardanega, 2016; William, 2024). Furthermore, underlined is the need for online academic integrity mastery training in improving students’ understanding of plagiarism and their opinions on academic dishonesty (Curtis et al., 2013).

Educators play a crucial role in shifting the perception of AI from a shortcut to a useful tool. They achieve this by instructing students on how to use AI tools, like automatic text summarization, and emphasizing the importance of using credible sources (Simpson, 2023). By combining proactive education and rigorous policy execution, academic integrity can be preserved in the AI era, ensuring that technological advancements enhance rather than undermine scholarly communication (Goel and Nelson, 2024).

With the spread of AI applications in education, concerns about the privacy and security of student data are increasing (Huang, 2023). The integration of AI in education offers opportunities for personalized learning experiences and advanced teaching methods (Eden et al., 2024; Wang et al., 2023). AI systems gather and evaluate students’ personal data, academic achievements, and learning patterns. This integration also gives rise to substantial apprehensions regarding data privacy and security (Sontan and Samuel, 2024; Alrayes et al., 2024; Li Y. et al., 2024; Risang Baskara, 2023). These worries encompass the spreading of personal information, data recognition, and breaches of security, particularly when it comes to handling data related to students (Cahyanto, 2023). Unauthorized access could infringe upon students’ privacy and result in potentially detrimental outcomes (Devi et al., 2023; Guan et al., 2023).

The ethical risks posed by AI technology to student personal information require robust security measures and clear communication regarding data processing to ensure informed consent (Devi et al., 2023; Huang, 2023). Current privacy policies are still limited, and there are many challenges to be faced in the use of AI and e-learning, especially facial recognition and automatic decision-making (Cahyanto, 2023). Therefore, addressing these privacy and data security concerns requires a comprehensive approach that includes establishing robust ethical frameworks, increasing transparency, and ensuring equitable resource allocation to close gaps in AI education practices (Ma and Jiang, 2023).

Thus, educational institutions must establish comprehensive data protection policies, use data transparently and responsibly, and protect students’ data privacy rights (Meszaros and Ho, 2021). AI-based education tools are used to lower these risks. It is important to think about being fair, ethical, and transparent (Harry, 2023; Nikolopoulou, 2024). Frameworks like Privacy by Design and Privacy Enhancing Technologies (PETs) give instructions on how to build privacy defenses into AI systems ahead of time (Sontan and Samuel, 2024). Also, making sure everyone has the same access to new technologies and fixing biases in AI algorithms are very important things to think about in educational AI applications (Sarwar et al., 2024). The smart use of AI in education can get the most out of these technologies while still protecting people’s privacy and data safety. It’s important to find a balance between the two.

Cultural factors in accepting AI in higher education

Factors influencing the acceptance of AI are as varied as they are with other educational technologies (Ismatullaev and Kim, 2024; Ye et al., 2019). Research has shown that sociodemographic factors can influence attitudes toward AI. The different patterns observed across factors such as country, age, and gender emphasize the importance of cultural context in shaping future perspectives on AI (Grassini and Ree, 2023).

Cultural factors play an important role in the acceptance of technological tools. For instance, Keller (2009) found that cultural factors significantly influence the acceptance of virtual learning environments. Positive or negative cultural influences can significantly affect the adoption of AI in educational settings. When looking at how AI is used in academia, it’s important to think about the things that affect people’s thoughts and plans about using AI, like how dangerous they think it is, how easy it is to use, and how much work they think it will take (Jain et al., 2022).

The Unified Theory of Acceptance and Use of Technology (UTAUT) and the perceived risk theory (Wu et al., 2022) both have many factors that affect how people feel about AI-supported learning environments. The elements encompass students’ interaction with AI, their risk perception, and the level of support provided by the learning environment. Furthermore, the UTAUT model links the cultural dimension to the social norm variable (Ismatullaev and Kim, 2024).

Perceived utility and user-friendliness influence the adoption of AI applications in Pakistan. The sociocultural elements have an impact on the views of university students toward the unconscious data collection carried out by AI technologies. Men, individuals with high incomes, and those majoring in business demonstrate lower levels of concern about such activities (Bokhari and Myeong, 2023).

Cultural factors play an important role in understanding the impact of AI on its acceptance in higher education. For example, in Asian cultures, historical context, religious beliefs, and levels of exposure to technology may contribute to a higher level of acceptance toward AI. Countries like Japan have embraced AI-based solutions extensively to serve an elderly population, and cultural values have been essential in this process. On the other hand, in the West, individualism and cautious views on technology can hinder the acceptance of AI. These variations suggest that culture significantly influences the acceptance of technological breakthroughs (Na et al., 2023; Yam et al., 2023).

The integration of AI into sectors including public policies and higher education can depend much on cultural norms, especially trust, openness, and transparency. For instance, these principles serve as a major guide for policy development in Scandinavian nations to guarantee the harmonic integration of AI into society. Likewise, in Japanese society, collectivist ideas have helped public services like health and education to embrace AI. But cultural differences—ethical and privacy issues, among others—may hinder these activities (Liu, 2023; Robinson, 2020). The adoption of AI in higher education calls for not only technological advantages but also awareness of the cultural surrounds and consumer values. Policies that take cultural differences into account can help make this process less difficult and lead to more people accepting AI (Ho et al., 2023; Neumann et al., 2024) the more it is used.

In the healthcare sector, cultural perspectives significantly influence AI adoption. High uncertainty avoidance poses a significant barrier. Clinicians’ preference for face-to-face interactions over AI-based ones shows how cultural factors like avoiding uncertainty can make it harder for people to accept new technology (Krishnamoorthy et al., 2022). A complex interplay of cultural factors is evident in AI acceptance. Culturally sensitive approaches are required to improve the adoption and integration of AI technologies across different regions and sectors.

In conclusion, the literature review highlights the increasing integration of AI in higher education, transforming teaching, learning, and academic communication practices. AI tools offer benefits such as personalized learning experiences, improved student engagement, and enhanced efficiency in administrative tasks. However, the adoption of AI also raises important ethical issues, especially around academic integrity, plagiarism, and data privacy. The impact of AI on scholarly communication is evident through changes in scholarly publishing, collaborative learning environments, and the quality of student interactions. Cultural factors have a significant impact on attitudes and acceptance of AI in educational settings. At the same time, these factors demonstrate differences between countries and conditions in which they are implemented. In light of advances in AI technology, educational institutions need to establish comprehensive regulations, protocols, and training initiatives to ensure the ethical and responsible use of these technologies. At the same time, it becomes imperative to maximize their potential to improve the educational experience in this context. We need more research to comprehensively understand the interplay between AI, academic discourse, and cultural influences in higher education.

Research methodology

We asked students attending universities in Turkey and Russia about their thoughts and experiences regarding the application of AI in academic communication. The purpose of this study was to analyze these views and experiences. The study employed a qualitative research design that aligned with its objectives. According to Creswell and Creswell (2017), qualitative research is a beneficial way to investigate complicated social issues because it gives a full picture of how people think, feel, and act in their own cultural setting. The utilization of the qualitative technique will enable the acquisition of comprehensive and intricate insights into participants’ perspectives on AI at universities. Furthermore, its potential to discern cultural disparities among Turkish and Russian students led to its selection. The study employed a comparative case study design, which entails methodically comparing two or more examples to discern similarities, differences, and trends (Yin, 2018). The participants in this study consisted of university students from Turkey and Russia. The comparison mostly centered around their viewpoints and encounters regarding AI in the realm of academic communication.

Participants

The study participants were 71 university students, consisting of 33 Turkish students and 38 Russian students. Students in Russia study at a state university in Russia. Students in Turkey study at a state university in one of the southeastern provinces of Turkey. We recruited the participants using a purposive sampling technique. The inclusion criteria for participants included their current enrollment in a university program in either Turkey or Russia, as well as their willingness to participate in the study. While the female rate among participants in the Russian group is 76.3%, it is 78.8% in the Turkish group (Table 1). Both university populations have females higher than male.

Table 1
www.frontiersin.org

Table 1. Demographic distribution of participants.

The primary criteria for sample selection were active university enrollment and voluntary participation in the study. Voluntary participation encouraged participants to share their experiences more openly, resulting in rich data. An appropriate foundation for cross-cultural comparison was the choice of public colleges in both nations and gathering similar numbers of data: 33 participants from Turkey and 38 participants from Russia. In addition, the similar gender distribution in both countries (Turkey: 78.8% female, Russia: 76.3% female) strengthened the comparability between the groups. Due to the qualitative nature of the study, it was aimed at capturing diversity of experience rather than statistical representation, and for this purpose, students from different academic departments were included in the study. However, it should be recognized as a limitation of the study that the voluntary participation principle may have encouraged the participation of students more interested in AI. Future studies can overcome this limitation by implementing a more systematic sampling strategy. In addition, future researchers can increase the diversity of participants by using a stratified sampling strategy. Since the nature of the study involves current technologies, future researchers can consider the level of familiarity and use of these technologies as a variable in participant selection.

As the authors, we acknowledge the limited sample size in this study for generalization purposes. However, the primary objective of this research is to provide an in-depth understanding, not to make statistical generalizations. The sample size is sufficient for the thematic analysis approach employed (Braun and Clarke, 2006). Future researchers and readers should consider the sample limitation when interpreting the study’s findings. In this context, future studies could expand on this research by using larger and more diverse samples. Also, the inclusion of a similar number of participants from the two countries (Turkey: 33, Russia: 38) provided a suitable basis for cross-cultural comparison. The data collected through open-ended questions enabled detailed information to be obtained from each participant, and only 242 codes could be made under the theme of “General Perceptions and Experiences.” Similar qualitative and intercultural comparative studies in the literature indicate that 30 to 40 participants from each group are acceptable. In addition, considering the difficulty of collecting data from two different countries and in two different languages, a sample of 71 students from public universities in both countries provided comparable and valuable data. We recommend testing the findings with larger samples in future studies.

Data collection tool

The present study utilized a survey with open-ended questions as the data collection instrument. We created the survey to allow students to share their perspectives and experiences with AI in academic communication. Three primary components divide the survey, each focusing on a specific aspect of the study inquiries.

Section 1: general perceptions and experiences

This section aimed to explore the participants’ general perceptions and experiences regarding the use of AI in academic communication. This section included the following open-ended questions:

1. What is your overall understanding and experience with using generative AI tools for academic communication?

2. How do you think the integration of generative AI in academic communication will impact your learning experience?

3. Can you envision any specific scenarios where generative AI tools could be particularly beneficial in your academic field or discipline?

Section 2: effects on academic communication and collaborations

This section focused on the perceived effects of AI-generated content on academic communication and collaborations. This section included the following open-ended questions:

1. Can you describe a scenario where AI-generated content could facilitate more effective communication among team members?

2. Can you describe any specific experiences or examples where generative AI has already impacted academic communication, either positively or negatively?

Section 3: evaluation of AI-generated content

The third section aimed to explore the participants’ evaluation of AI-generated content in terms of academic integrity and plagiarism. This section included the following open-ended questions:

1. What concerns do you have about the use of AI-generated content in academic settings?

2. Can you describe a situation where AI-generated content might not be suitable or appropriate for academic purposes?

3. How do you approach the issue of academic integrity and plagiarism when using generative AI for assignments or papers?

4. How comfortable would you feel submitting work that was partially generated by AI to your professors? Explain your answer.

The research team developed the open-ended survey through a comprehensive review of existing literature on AI in higher education and academic communication. Three faculty members from each country then examined the questionnaire questions and obtained expert opinions for content validity. The experts confirmed that the questions were appropriate. We translated the questionnaires into Russian and Turkish, and then administered them to the students in their native languages. Next, we conducted a pilot study with five students from each country. We asked the students to think aloud as they answered the questions in the pilot study, identifying the points they did not understand. We provided clear instructions to the participants and assured them of the confidentiality and anonymity of their responses. The students understood that their participation was entirely voluntary. We made it clear to the students that they could leave the study at any point during the research process. Additionally, we recorded the answer time for each question and optimized the questionnaire’s completion time. As a result of all these processes, the questionnaire was finalized. This comprehensive validation process shows that the data collection tool is appropriate and reliable to answer the research questions.

Due to the nature of the open-ended questionnaire, a qualitative approach was used in the study instead of quantifying student responses. Collecting opinions and other documents from students, in addition to their behaviors or studies, can achieve data triangulation. In addition, quantitative scales can be developed to examine their opinions and attitudes. In fact, the data obtained in this questionnaire has the potential to form the basis for a quantitative scale. In this context, it is suggested that future researchers should differentiate the data collection tools.

Data analysis

Thirty-three Turkish and thirty-eight Russian university students’ qualitative material was examined using Braun and Clarke's (2006) six-stage theme analysis system. All researchers participated in the rigorous and tight approach used in the analytical procedure to guarantee the quality and confidence of the outcomes.

Stage 1: familiarizing with the data

Every survey response was translated into English. Two separate translators—Turkish-English and Russian-English—reviewed the translational correctness in order to ensure semantic equality. The study team then combed over the translated responses again several times in order to spot the first likely trends and interpretations in the data.

Phase 2: initial coding

Using an inductive technique, all researchers separately initial coded the data. Under this method, codes sprang from the material instead of being predefined. Every researcher allocated initial codes and methodically went over the whole data set looking for significant language passages. The coding was very detailed in order to analyze cultural richness and differentiation. For instance, this process generated 242 codes solely under the theme of “General Perceptions and Experiences.”

Inter-coder reliability and consensus process

Three researchers independently coded the entire data set and produced 728, 705, and 710 codes, respectively. Inter-coder reliability was calculated using Miles and Huberman’s (1994) formula. Reliability scores between pairs of researchers are as follows:

• Between the first and second researcher: 95.39% (33 code discrepancies).

• Between the first and third researcher: 94.85% (37 code discrepancies).

• Between the second and third researcher: 96.04% (28 code discrepancies).

We computed an average intercoder dependability of 95.43%. This rate demonstrates great coding consistency and is much above the 80% level advised by Miles and Huberman (1994). Consensus sessions, when the researchers addressed the justification for their coding choices, helped settle differences. The researchers developed the final coding system once they resolved all contentious rules.

Stage 3: searching for themes

The research team then cooperatively examined the initial codes in search of possible trends. This approach arranged the codes according to their correlations and trends into more general themes. The team investigated the relationship between codes and possible themes.

Phase 4: reviewing themes

To ensure the validity of the themes, the research team conducted two levels of analysis:

• Level 1: Comparison of themes with coded quotations.

• Level 2: Examining the relationship of themes with the whole data set.

At this stage, the researchers adjusted, merged, or separated the themes as necessary. The researchers paid particular attention to ensuring that each theme was consistent and distinctive.

Phase 5: identifying and naming themes

Every theme was precisely defined by the team to faithfully depict the data it stands for. The main idea of every subject was precisely expressed by the chosen, succinct, but instructive theme names.

The theoretical framework in data analysis

For understanding the results, the Unified Theory of Acceptance and Use of Technology (UTAUT) model and the theory of perceived risk were used. These theories talk about how people accept and use technology. We specifically used “social norm,” a sub-dimension of the UTAUT model, as a powerful tool to interpret the data. Social norm refers to the social expectations that shape individuals’ behaviors and therefore serve as an important guide in understanding individuals’ preferences and attitudes towards technology use. Social norm is one of the basic building blocks of the UTAUT model and explains the influence of the social environment on individuals’ decisions about technology use. We evaluated the participant’s statements using the relevant codes in our framework. On the other hand, perceived benefit within the model is an important element affecting the use of AI. Specific aspects of “Performance Expectancy” focus specifically on the potential of AI to accelerate learning processes, facilitate access to information, and increase academic achievement. These aspects contribute to users developing positive attitudes toward AI technology. In this context, we evaluated the positive and negative statements students made about the use of AI in academic communication. For example, statements such as “Facilitating research” and “Better presentation” were categorized as positive, while statements such as “Incorrect and inconsistent information” were categorized as negative. This evaluation provided a detailed understanding of the participants’ perceptions of AI.

Another theoretical framework is Hofstede’s cultural dimensions theory. The findings were analyzed in the context of the individualism–collectivism and uncertainty avoidance dimensions of this model. For example, in the individualism–collectivism dimension, students’ preference to use AI in collaborative projects was associated with collectivism, while their preference to use it in individual tasks was associated with individualism. In the uncertainty avoidance dimension, it was observed that concerns about privacy and data security increased this tendency. For instance, we found an association between the individualism–collectivism dimension and students’ preference for collaborative or individual work when using AI. On the other hand, codes that mentioned concerns about privacy and data security were associated with the uncertainty avoidance dimension.

Both theoretical frameworks are quite suitable for making sense of the study’s data. While the UTAUT model provides a strong theoretical basis for analyzing students’ behaviors of accepting and using AI technology (Acosta-Enriquez et al., 2024), it was effective in explaining the differences in students’ attitudes towards AI with its dimensions such as performance expectancy and effort expectancy. On the other hand, Hofstede’s Cultural Dimensions Theory provides an appropriate framework to make sense of the differences in the use of AI in two different cultural contexts. When these two theoretical frameworks were put together, they gave a full picture of how technology acceptance and cultural differences affect the use of AI.

Future researchers can analyze these data using different theoretical frameworks and approaches. Theoretical frameworks serve as a guide to making sense of and explaining the data, helping researchers to understand complex concepts more clearly. These frameworks can be considered a powerful analysis tool that illuminates different aspects of the data. Analyzing from different perspectives can naturally lead to different results.

Quality assurance measures

We employed several strategies to enhance the analysis’s reliability. The researchers held regular online meetings to discuss and refine the coding process. We documented all decisions made throughout the process in detail and created an audit trail to ensure transparency. Member checking was conducted with a subset of participants to confirm the accuracy of the interpretations obtained. In addition, peer review meetings were organized with external researchers to strengthen the objectivity of the study by questioning assumptions and interpretations in the research process.

While still sensitive to the subtleties of how Turkish and Russian students view and experience AI in academic settings, the study concentrated on spotting trends both inside and between the two cultural settings. Throughout the research process, the team stayed conscious of possible biases and often spoke about how they could influence data interpretation.

Results

The General Perceptions and Experiences theme looks at students’ overall views of AI tools, including their perceived usefulness, position as assistance in academic duties, and related concerns. It also points up different degrees of student AI acceptance. The focus of our study is to analyze the influence of AI on student interactions and collaborative work in the context of academic communication and collaborations. The analysis contrasts the positive effects, like better teamwork within groups, with the negative perceptions, like worries about less strong analytical thinking. The evaluation focuses on the ethical issues of using AI in academia, specifically in relation to academic integrity and plagiarism. The study covers students’ degree of comfort with AI-generated content, plagiarism concerns, problems with content quality, and more general ethical issues. This study examines the perspectives of Turkish and Russian students, emphasizing the cultural and educational disparities in their comprehension and application of AI. In the study, code diversity was kept as high as possible in order to observe cultural differences more easily (see Figure 1).

Figure 1
www.frontiersin.org

Figure 1. Concept map of themes and categories.

General perceptions and experiences regarding the use of AI in communication

The category of attributes of AI tools relates to the way respondents characterize AI depending on their experiences. Respondents from Turkey and Russia noted several advantages of AI tools (Russia: 20, Turkey: 19). Both nations’ participants rank AI tools as “Useful” (Russia: 17, Turkey: 21) and “Beneficial” (Russia: 4, Turkey: 6). Whereas a Turkish student said, “I’ve used it before and it’s quite useful,” (T_16), one of the Russian students said, “It’s a very good and useful application” (R_19). Some students cited “Easily” (Russia: 6, Turkey: 4), “Effective” (Russia: 4, Turkey: 4), “Faster” (Russia: 4, Turkey: 4) and “Helpful” (Russia: 4, Turkey: 2). While another student from Turkey said, “I see it as a good and effective shortcut for my research,” and a Russian student said, “It helps in the preliminary preparation stage” (R_10). Another student from Turkey stated, “AI analyzes data sets quite effectively.” For us students in writing and editing articles, it especially offers ease; another student commented, “…I got help while doing my homework.” Furthermore, some individuals called AI technologies “Productive” (Turkey: 1) and “Successful” (Russia: 1). These results imply that people in both nations value and know the beneficial aspects of AI tools (see Table 2).

Table 2
www.frontiersin.org

Table 2. Code frequencies for general perception theme.

The findings in the “AI Tools’ Feature” category show that students’ general evaluations of AI are largely positive. Especially the high frequency of the code “useful” (Russia: 17, Turkey: 21) reveals that students in both countries see AI as a practical tool in their academic studies. Still, what’s fascinating here is that Turkish students emphasize AI’s importance more often. This difference can be explained by the greater readiness of Turkish students to adapt to new technologies or by the greater openness of the Turkish educational system toward the integration of technology.

Respondents’ opinions about how AI products simplify their lives or tasks determine the category of AI’s assisting function. Turkey’s respondents emphasized the benefits of AI more than those of Russia (Turkey: 43, Russia: 27). AI can be beneficial in the following areas: “Finding information and source” (15), “Doing homework” (5), “Investigating unknown topics” (1), “Saving time” (2), and “writing thesis and academic works” (3), according to respondents from Turkey. Participants from Turkey stated, “I generally use it to find information that I cannot access otherwise. It’s quite useful for obtaining organized information for courses” (T_20), “I sometimes used it while making slides for my courses” (T_10) and “I view it positively and use it myself in terms of resources” (B_30). It was also stated that AI can “Expand our world of thought” (2), “Idea generation” (1), “Correcting our writing mistakes” (1), “Creative writing” (1), “Data analysis” (1), “Helping attention and concentration” (1), “Helping daily routine” (1), “Increase logical thinking” (1), “Providing quality education for all” (1), “Reducing the human workload” (1) and “Solving problems” (1). Participants underlined in Russia that AI can support “Assignment” (9), “Finding information and source” (7), “Exploring unknown topic” (5), “Time-saving” (3). Some student comments on the matter include “Generally useful for my homework and research. (R_36)” and “In my experience, ChatGPT can answer every question I ask. I have no trouble finding responses to my questions. (R_25).” These results imply that participants—especially in Turkey—believe that AI can play several beneficial functions in scholarly communication.

When the “Assistant Role of AI” category is analyzed, it is seen that Turkish students (43) use AI for more diverse purposes than Russian students (27). In particular, the fact that Turkish students used the code “finding information and source” more often (15 times vs. 7) suggests that they use AI more as a tool for searching for information and doing research. Conversely, Russian students would rather employ AI for particular chores, including homework (assignment: 9). The variations in the research-assignment balance between the two countries’ educational systems could help to explain this disparity.

The negative effects and concerns category captures students’ worries and negative experiences with AI. Some participants in both nations expressed worries about the negative effects of AI applications. Participants in Russia rated AI as “Nonpermanent-Knowledge” (4), “Not Be Sufficient” (2), “Not Effective” (2), “Should Be Improved” (2), “Eliminate Our Creativity” (1), “Incorrect Information” (1), “Damages the desire for research and learning” (1) and “reducing the student’s thinking skills” (1). About this matter, they expressed, “However, it’s also an application that needs improvement, as it sometimes cannot answer questions. (R_24)” and “It will not be permanent because it’s readily available information. (R_15).” Participants assessed AI’s “Reduce Discipline” (2), “Addition” (1), “Make People Lazy” (1), “Reducing the quality of learning” (1) and “Rote Learning” (1) in Turkey. One of the Turkish students, T_17, said, “…may reduce the quality of learning”; the other student, T_33, said, “I think it will lead to rote learning.” These results imply that certain respondents from both nations have unfavorable opinions and worries about the application of AI.

In the “Negative Effects and Concern” category, Russian students (17) expressed more concerns than Turkish students (13). In particular, the fact that the code “nonpermanent-knowledge” was seen only in Russian students (4) reflects the importance that the Russian education system attaches to permanent learning. Turkish students’ concerns, on the other hand, focus more on behavioral issues such as loss of discipline (“reduce discipline”: 2).

In both Russia and Turkey, an equal number of respondents (13) indicated they lacked knowledge of AI techniques. This result implies that some respondents in both nations have not yet used AI tools or have just limited knowledge of them. This finding is unsurprising given the relatively recent emergence of AI applications in academic communication.

Overall, it seems that respondents in Russia and Turkey view and experience AI tools in somewhat different ways, albeit with certain parallels. Respondents in both nations value the advantages of AI tools and believe they can be rather useful for academic communication in several ways. While Turkish respondents concentrated more on the beneficial usage of AI, Russian respondents gave the negative effects and worries some more weight. Furthermore, some respondents in both nations appear to have never used AI tools. These results imply that although opinions and experiences of the use of AI technologies in academic communication may vary between countries, generally there is awareness of the possible effects of these tools.

Effects of AI-generated content on academic communication and collaborations

We generated 16 codes related to the positive effects of using AI on academic communication. Participants in Turkey mentioned the positive impact of AI on academic communication more than those in Russia (Turkey: 31, Russia: 18). In Turkey, codes such as “group working” (8), “positive communication” (7), “better presentation” (4) and “increase interaction” (3) came to the fore. Students said in this regard that AI tools have good contributions in communication with the words “It strengthens communication among classmates (T_11)” and “It helps to fill our knowledge gaps and makes us more confident in communication. (T_20).” In Russia, the codes “facilitates communication” (3), “increases communication” (2) and “facilitating research” (2) were stressed more often. Russian students reported on the other hand, “an introverted individual with communication issues might try to communicate with friends using an AI-generated scenario example. (R_33)” and “it facilitates communication (R_27).” These findings suggest that mostly Turkish participants feel AI could increase contact and cooperation in academic communication.

Positive perceptions represent the positive statements used by students regarding AI tools. Participants in both countries have positive perceptions about AI (Russia: 24, Turkey: 22). “Access information” (Russia: 10, Turkey: 9) and “speed” (Russia: 2, Turkey: 4) are the most frequently mentioned positive perceptions in both countries. In this context, a Russian student stated that “Accessing more resources can lead to more effective communication. (R_14)” whereas a Turkish student said, “we can access information more easily, quickly, and reliably. (T_04).” Furthermore, whereas in Turkey it is “useful” (2) and “generating ideas” (2), in Russia participants underlined that AI is “easier” (3) and “positively” (3). Whereas the Turkish student said, “…having a language control tool is very useful for students studying in English. (T_14),” the Russian student said, “as I mentioned in my previous answers, [AI is] easy and practical. (R_19).” These results reveal that people in both nations understand the benefits of AI in scholarly communication.

We collected eleven different codes related to students’ negative perceptions of AI. Participants in Turkey had more negative perceptions about AI than those in Russia (Turkey: 20, Russia: 11). In both countries, “incorrect and inconsistent information” (Russia: 2, Turkey: 4), “ready-made solutions” (Russia: 1, Turkey: 4) and “reduce our thinking and research ability” (Russia: 2, Turkey: 3) were the most frequently mentioned negative perceptions. In this context, a Turkish student used the phrase “…producing misleading content… (T_13)” and similarly a Russian student used the phrase “it can provide incorrect and inconsistent information. (R_38).” Furthermore, while participants in Turkey stated that AI is “making people lazy” (2), “limits the individual’s creativity” (2) and “monotony” (1), participants in Russia stated that AI is “not permanent” (1) and “get in the way of people’s work” (1). One Turkish student (T_20) said “…making people lazy and leading them to incorrect information.” Also, Russian student R_08 said that “it can get in the way of people’s work.” These results imply that participants—especially in Turkey—have worries about AI possibly having some detrimental consequences on academic communication.

We identified seven codes under the category of negative effects of using AI in academic communication. In this context, the negative effects of AI on academic communication were mentioned by participants in Russia more frequently than in Turkey (Russia: 10, Turkey: 4). The most frequently mentioned negative impacts in both countries were “unfairness” (Russia: 3, Turkey: 1) and “reduce communication” (Russia: 2, Turkey: 2). A Russian student (R_04) stated “I do not think so; on the contrary, it might reduce communication,” while a Turkish student (T_06) similarly stated “I think such applications could negatively affect my interaction with classmates rather than increase it.” In addition, while in Russia participants emphasized issues such as “plagiarism” (2), “cheating” (1) and “personal information” (1), in Turkey “academic ethical issues” (1) were mentioned. Russian student (R_36) used the expression “it can lead to plagiarism.” The Turkish student (T_13) used the phrase “…producing misleading content and academic ethical issues.” These findings show that participants in both countries are aware that AI can lead to some negative consequences in academic communication.

Respondents from Russia claimed more lack of knowledge about AI than those from Turkey (Russia: 14, Turkey: 8). This result implies that users of AI in academic communication have poorer knowledge among Russian respondents.

Regarding the effects of AI on academic communication, it appears that the respondents in Turkey and Russia have certain parallels as well as variances. Turkish people worry about AI’s negative effects, but they also think it could improve academic collaboration. Russian players, on the other hand, talked about both the pros and cons of AI. People in both countries know that AI can help with academic communication, but they also think that it can lead to “incorrect and inconsistent information,” “ready-made solutions,” and “lessen our ability to think and do research.” Also, people in Russia seem to have less experience with AI. These findings indicate that perceptions and approaches to AI use in scholarly communication vary across different countries. Both countries should think about the pros and cons of this use of AI (see Figure 2).

Figure 2
www.frontiersin.org

Figure 2. Relation map on academic communication and collaborations based on countries.

When we look at the findings of Table 3 with a focus on cultural context and communication, significant differences emerge. While Turkish students tend to see AI as a tool for group work and communication (Group work - Turkey: 8, Russia: 2), Russian students focus more on individual use. This reflects the collective work tendency in Turkish culture. A similar difference is seen in the perception of communication; Turkish students emphasized more on “positive communication” (7) and “increase interaction” (3), while Russian students emphasized “facilitates communication” (3). Although students from both countries value access to information (Russia: 10, Turkey: 9), Turkish students are more inclined to share and discuss this information. Cultural differences are also evident in negative perceptions. Turkish students are more concerned about ready-made solutions (4), while Russian students are more worried about reduced communication (2). The fact that inexperience is more prevalent among Russian students (14), while Turkish students use AI tools more actively (8) shows cultural differences in technology adaptation. In terms of ethical concerns, Russian students emphasized unfairness and plagiarism more, while Turkish students had a broader view of academic ethics. In terms of communication styles, Turkish students preferred presentation and interaction, while Russian students preferred direct communication, reflecting differences in cultural communication styles. Cultural values influence the use of AI in academic communication, as all these findings demonstrate.

Table 3
www.frontiersin.org

Table 3. Categories and codes related to AI-generated content on academic communication and collaborations.

Evaluation of AI-generated content for academic integrity and plagiarism

Participants in both nations perceive AI-generated content similarly: 37 in Russia and 33 in Turkey. Still, participants in Turkey feel more “comfortable” (Turkey: 8, Russia: 6) and “comfortable with my contribution” (Turkey: 9, Russia: 5) providing AI content. Russian student R_21: “I present the data after checking their source. I feel comfortable if it is precise and accurate,” and Turkish student (T_10) “If I have prepared it really consciously and made honest use of AI, I can present it with confidence.” Russia’s participants reported they were “partially comfortable” (5). Most of the participants in both nations said they were “uncomfortable” (Russia: 19, Turkey: 16) with seeing AI stuff. In this regard, a student from Turkey (T_15) said “I feel uncomfortable, I think I tricked the lecturer,” while a student from Russia (R_03) said “I am not comfortable that our entries are inspected in general.”

Respondents in Turkey mentioned academic integrity issues more often than those in Russia (Turkey: 39, Russia: 30). In both countries, “plagiarism” (Russia: 11, Turkey: 15) was the most frequently mentioned problem. For example, a student from Turkey (T_24) stated “There is absolutely no honesty in taking advantage of the whole way; it is also plagiarism.,” while a student from Russia (R_03) stated “Plagiarism, it’s quite common.” In Turkey, the participants also emphasized issues such as “be honest” (6), “not appropriate for academic assignments” (4) and “need policies and rules” (3), while in Russia issues such as “non-honest” (3) and “lack of effort” (3) were mentioned. In this context, Turkish student (T_15) stated “Even though I try not to do it, many people of our generation do it and violate the rights of other students, and I think there should be some rules to prevent this.” The Russian student (R_30) used the expression “need to be honest and academically oriented.”

Participants in Turkey mentioned content quality issues more than those in Russia (Turkey: 22, Russia: 16). While “inconsistencies, and missing information” (11) was the most frequently mentioned problem in Turkey, “ready-made information” (6) came to the fore in Russia. Problems such as “monotony” (Russia: 1, Turkey: 2) and “unproductive” (Russia: 2, Turkey: 1) were expressed in both countries. In this context, the Russian student (R_07) said, “[AI] can combine unclear and meaningless arguments, which leads to an inconsistent and poor desired outcome.” While Turkish student (T_33) used the expression “I think it is not accurate information.”

Participants in both nations expressed comparable worries regarding ethical and legal aspects of AI use (Russia: 8, Turkey: 8). Russia highlighted “privacy and data concerns” (4) more frequently, while Turkey highlighted “ethical issues” (4). In this regard, a Turkish student (T_22) said “Use of AI in important assignments or presentations is not appropriate as it is unethical.” One Russian student (R_13) said “We may experience some privacy issues.” Both Russia and Turkey voiced worries about “use for malicious purposes” (Russia: 2, Turkey: 1).

Respondents in Russia hold a more pessimistic perception of AI in comparison to those in Turkey (Russia: 22, Turkey: 16). The most cited unfavorable opinions in both Russia and Turkey are “laziness” (Russia: 6, Turkey: 6) and “losing profession” (Russia: 3, Turkey: 3). One Russian student (R_23), for instance, said “It increases the tendency to be lazy.” A Turkish student (T_25) said “I think artificial intelligence tools can overtake manpower.” While in Turkey, issues such as “hinder critical thinking” (2) and “accepting without critiques” (2) were stressed, participants in Russia also stated opinions such as “negatively” (4) and “restricts our research ability.” Accordingly, a student from Turkey (T_23) said “Loss of critical or creative thinking in users,” while a student from Russia (R_05) said “Excessive use negatively affects a person’s own learning.”

Respondents in Russia exhibited marginally more favorable attitudes toward AI in comparison to those in Turkey (Russia: 18, Turkey: 15). In Turkey, the most commonly expressed favorable opinions were “get support” and “acceptable,” while in Russia, “no concern” and “positive” were the most prominent. In this context, a Russian student known as R_22 stated that they currently have no concerns or anxieties. The Turkish student (B_13) expressed that while AI can offer assistance, they possess the authority to make judgments and assess the outcomes.

Participants in both countries think that the suitability of AI for academic purposes depends on the context (Russia: 12, Turkey: 10). In Russia, “suitable” (5) and “suitable for academic purposes” (4) were emphasized more frequently, while in Turkey, opinions such as “be sufficient for objective issues” (2) and “depends on purposes” (2) were expressed. In this regard, the Russian student stated “It may be appropriate if we offer different ideas.” Turkish student T_02 stated, “It’s crucial to consider the intended use of this content.” A presentation for a lecture is appropriate, while a paper or article is not.”

In both countries, participants mentioned the impact of AI on the learning process (Russia: 8, Turkey: 9). “Access information” (Russia: 4, Turkey: 5) and “easier” (Russia: 1, Turkey: 2) are positive impacts mentioned in both countries. However, participants in Russia were more likely to mention “no-knowledge” (8). T_03, a Turkish student, expressed that the platform is highly convenient for efficiently retrieving material from previous studies as well as fresh findings. A Russian student (R_17) observed, “It is beneficial for us when we need to compose a summary or retrieve information…”

Regarding cultural background, the findings in Table 4 reveal very significant differences between the two nations in the way AI material is presented. While Turkish students are more at ease presenting materials (Comfortable: Turkey 17, Russia 13), Russian students are just somewhat comfortable—Partially Comfortable: Russia 5, Turkey 0. In the perception of academic honesty, it is noteworthy that Turkish students emphasize this issue more (Turkey 39, Russia 30) and especially the issue of being honest (Be honest: Turkey 6, Russia 2). Cultural differences are also evident in terms of content quality, with Turkish students more concerned about inconsistency and incomplete information (Turkey 11, Russia 3), while Russian students see the use of ready-made information as more problematic (Russia 6, Turkey 4). We observe a similar divergence in ethical and legal concerns. Russian students were more concerned about privacy (privacy concerns: Russia 4, Turkey 2), while Turkish students were more worried about ethical issues (ethical issues: Turkey 4, Russia 1). Both groups shared a common concern of laziness, with six students in each group, while Russian students were more concerned about their research abilities being limited. These differences can be considered a reflection of the Turkish educational system’s emphasis on group work and sharing and the Russian education system’s emphasis on individual work and originality. In addition, these findings also reveal differences in cultural attitudes toward technology adaptation, the concept of academic honesty, and data security.

Table 4
www.frontiersin.org

Table 4. Categories and codes on evaluation of AI-generated content for academic integrity and plagiarism.

We observed significant cultural differences in the attitudes and experiences of Turkish and Russian students toward the use of AI. Turkish students tend to see AI as a means of group work and communication. This reflects the collectivist nature of Turkish culture. Information sharing and discussion are more welcome among Turkish students. Their approach to using AI also demonstrates more leisure and entrepreneurship. Conversely, Russian students give more of their individual usage top priority. Their propensity to shun ambiguity drives them to be more wary of AI. Russian students are more concerned about data security and privacy issues. Academic honesty is an important issue in both cultures. However, while Turkish students look at the issue from a broader perspective, Russian students express more specific concerns. Turkish students’ more active use of AI and evaluation of it in various fields indicates a low uncertainty avoidance tendency. Russian students’ emphasis on individual learning and their cautious approach in the use of AI reflect high uncertainty avoidance tendencies. These findings are consistent with the collectivism–individualism and uncertainty avoidance dimensions of Hofstede’s cultural dimensions theory.

Discussion

Our study reveals complex and intriguing relationships between AI technology, academic communication, and culture in Turkish and Russian university settings. Students from these two countries have the same and different perceptions and ways they use AI tools in academic communication. It depends on culture and how people see technology.

General perceptions and experiences

In terms of general perceptions and experiences, students from both Turkey and Russia recognized the potential benefits of AI in increasing the accessibility of information and streamlining academic tasks. This finding is consistent with previous research highlighting the positive impact of AI on student learning experiences (Aldosari, 2020; Jaboob et al., 2024; Yan et al., 2024). According to the results of the study conducted by Tierney et al. (2025) and Stöhr et al. (2024) university students have more positive thoughts about the use of AI. From the perspective of the UTUAT model, it can be associated with the performance expectancy dimension. Acosta-Enriquez et al. (2024) and Sergeeva et al. (2025) demonstrate that the perceived performance expectancy of AI technology is a highly positive factor. The more positive students’ performance expectations, the more likely they are to continue using it. Turkish students have a more positive attitude about the collaborative potential of AI. The results (Kim and Lee, 2023) showed that student-AI collaboration has a big impact on creativity in content, expressivity in expression, and public utility in effectiveness. These effects changed based on how students felt about AI or how good they were at drawing. But Russian students placed more emphasis on privacy and data security concerns. Cultural differences in how people avoid uncertainty (Ismatullaev and Kim, 2024) and how they feel about using technology (Grassini and Ree, 2023; Sarwari et al., 2024) can explain these differences in focus.

Impact on academic communication and collaborations

The study’s findings demonstrate that AI exerts a substantial influence on academic communication and partnerships among university students. According to the studies (Hohenstein et al., 2023; Sarwari et al., 2024), AI and technologies based on it have sped up communication and the gathering of information, created new ways for people to interact, and especially helped people who are lonely connect with each other. Specifically, Turkish participants highlighted the favorable impacts of AI on collaborative tasks, interpersonal engagement, and the overall standard of communication. Previous studies support this finding. According to studies, AI helps to create a collaborative learning environment (Porter and Grippa, 2020; Son and Jin, 2024). Students prioritize its utilization as an assistant (e.g., reviewing and preparing for lectures, assisting with assignments), while educators utilize it as a content production assistant (e.g., personalizing content, writing lecture notes) (Clos and Chen, 2024). However, Russian students are more concerned about the negative impact of AI, such as less critical thinking and lack of originality. This assessment may be due to the cultural difference between how students learn and what they value in education (Keller, 2009). The performance expectancy dimension of the UTAUT model is directly related to these findings. Especially the positive statements emphasized by Turkish students indicate that they have high performance expectations. Russian students’ positive perceptions such as easy access to information can also be considered as one of the factors supporting performance expectancy. In this respect, both positive and negative attitudes of the students reveal the importance of the UTAUT model in student motivation and adoption behaviors.

Evaluation of AI-generated content

The findings of this study reveal that there are significant cultural differences in the evaluation of AI-generated content in terms of academic integrity and plagiarism. Participants in both Russia and Turkey highlighted the risks and opportunities of using AI in the academic environment, but these considerations are shaped by different priorities and approaches.

The findings show that respondents from Turkey approach the issue of academic integrity and plagiarism in the use of AI from a broader perspective. In particular, themes such as the emphasis on “being honest” and “the necessity of policies and rules” are prominent in Turkey. This is in line with Foltynek et al. (2023) or Karkoulian et al. (2024) discussions on the need to define the ethical framework of AI use. Goel and Nelson (2024) offer a similar perspective, arguing that the principles of academic integrity need to be reinterpreted in the face of emerging technologies. In Russia, academic integrity concerns are expressed in terms of the risks of slipping into “ready-madeness,” losing the ability of students to produce original thinking, and cheating. According to the results of the study by Stöhr et al. (2024), academic integrity and ethical issues stand out among students’ concerns about the use of AI. 61.9% of the students surveyed considered the use of AI-based chatbots in assignments and exams as a form of cheating. Moreover, a significant number of respondents (54.2%) expressed concerns about the potential impact of these technologies on future learning processes. Among students, these concerns were particularly centered on ethical concerns, whether AI tools undermine educational goals, and how they would impact academic achievement. However, students were generally opposed to the idea of banning the use of AI altogether and recognized that these tools can provide certain benefits in education.

In the study, Turkish students are more concerned about inconsistency and incomplete information from AI, while Russian students find the use of “ready-made information” more problematic. This reflects the differences in the understanding of learning in both countries. It can be said that Turkish students try to internalize the information presented by AI more in both group work and individual research, whereas Russian students see “ready-made” information as a threat to their own research process. Indeed, Cotton et al. (2024) make similar warnings that the practicality offered by technological tools may weaken student research skills.

Although ethical and legal concerns were raised in both countries, Russian participants focused more on “privacy and data security” concerns, while Turkish participants drew attention to the potential for “ethical violations.” This distinction reveals that, as Bokhari and Myeong (2023) emphasize, the approach to the use of technology may vary in the educational cultures and legislative structures of different countries. Both Turkish and Russian students share a common concern about “malicious use,” underscoring the need for social and academic spheres to monitor technological developments.

Analysis of the findings within the framework of Hofstede’s cultural dimensions theory reveals striking differences. Participants in Turkey tend to see AI as a tool for sharing and group work with a more “collective” attitude. Moreover, in Turkey, there is relatively less uncertainty avoidance towards the use of AI for different purposes, which enables participants to experience AI in a more relaxed and “entrepreneurial” way. In Russia, on the other hand, the emphasis on individual work is characterized by a higher sensitivity to uncertainty (high uncertainty avoidance) and prioritization of data security concerns. Thus, Russian students are more likely to use AI with caution, worry about losing their own research skills, and have privacy concerns. These results support Hofstede’s collectivism–individualism and uncertainty avoidance dimensions.

The most frequently expressed issue in both countries’ negative perceptions of AI is that it causes laziness. Additionally, Russian students emphasize concerns such as “limitation of research ability” and “negative impact,” while Turkish students emphasize the risks of “preventing critical thinking” and “accepting without questioning.” However, it is also noteworthy that “not feeling fear or anxiety” and “finding it positive” are more common in Russia. In Turkey, on the other hand, the perceptions of “receiving support” and “acceptability” are slightly more prominent. The nuances of positioning AI technologies in education and integrating them into personal learning practices reflect this dichotomy.

In both groups, it was observed that attitudes change according to the context in which AI is used. While students in Turkey see AI as a practical tool for homework and research, in Russia it can be found useful for purposes such as “summarizing” or “gathering information.” On the other hand, Russian students emphasize the risk of “no-knowledge” more and state that consuming information quickly may prevent deepening. This is in line with the view put forward by Song (2024) and Perkins and Roe (2024) that the effects of AI use on student learning outcomes should be investigated more extensively.

Participants emphasized the need for institutional arrangements and guidelines in both countries. As Fowler (2023) and Yusuf et al. (2024) point out, it is inevitable for higher education institutions to develop rules appropriate to the new conditions created by AI in education. These regulations should both protect academic integrity and encourage the utilization of the pedagogical benefits of AI. The findings suggest that in Turkey, these rules tend to be shaped in a more broadly participatory and collective framework, whereas in Russia they need to focus on individual responsibility and data privacy.

In conclusion, this research has shown that cultural differences in AI use are important in terms of academic integrity, content quality, ethical-legal issues and learning processes. Instead of a single “generic” AI policy, a more sustainable approach would be for each country to adapt within the framework of its own educational system and cultural values.

In this context, culture-specific approaches should be adopted in drafting AI policies and collectivist or individualist tendencies as well as uncertainty avoidance levels should be taken into account. Both in Russia and Turkey, programs should be developed to raise awareness about the concept of plagiarism and ethical use of AI tools. In particular, Russian students’ privacy concerns call for the implementation of stronger data protection measures on AI-based platforms. Turkish students have highlighted information inconsistencies and gaps, prompting the development of mechanisms like citation and automatic reference scanning.

Finally, the effects of AI tools on learning outcomes should be assessed, not only in terms of speed and access but also in terms of how they shape deep learning, critical thinking, and research skills. These recommendations are instructive for both institutions and policymakers. In the field of education, where AI technologies have become widespread, it is essential to develop comprehensive regulations and practices that take into account student experiences and cultural contexts. Maintaining academic integrity and ethical values will ensure effective utilization of the benefits offered by AI.

Cultural factors and AI adoption

The findings of this study make it clear that cultural elements are too important to be ignored when analyzing the implementation and outcomes of AI in higher education. Especially technological acceptance models, such as the Unified Theory of Technology Use and Acceptance (UTAUT), consider cultural elements as a key component (Venkatesh, 2000; Wu et al., 2022). In this model, the cultural dimension is closely related to the social norm variable, which helps explain in more depth the reasons why individuals adopt or resist new technologies. Indeed, based on perceived risk theory, Zheng (2017) examined the different effects of privacy concerns, visibility of security measures, and online seller reputation on user behavior in different cultures. This study reveals that understanding the components of trust and the way users embrace technology breakthroughs depends critically on a cross-cultural perspective. Comparably, Yusuf et al. (2024) discovered that cultural elements greatly influence participants’ favorable and negative opinions on generative artificial intelligence technology. Moreover, the results (Mansoor et al., 2024) revealed that participants’ opinions about artificial intelligence and their inclination to employ it vary depending on their demographic and intellectual background. Considered the most crucial factors were nationality and academic background. All these studies underline how important culture is for the adoption of artificial intelligence technology in higher education and cannot be regarded as a mere “background” feature. The attitudes of academics and students about new technologies can be shaped by cultural values, beliefs, and conventions; so, consideration to cultural variations in the design, distribution, and assessment of the results of applications is necessary. Therefore, it is important to properly utilize the possibilities of artificial intelligence in higher education through the interactions of cultural elements among several stakeholders and their influence on participants’ opinions of the technology.

When we evaluate our findings within the framework of the Unified Theory of Acceptance and Use of Technology (UTAUT), there are differences in two main dimensions. The first is performance expectation. Turkish students have a more positive view of the contribution of AI to their academic performance. For example, the differences in the codes of “group work” (Turkey: 8, Russia: 2) and “positive communication” (Turkey: 7, Russia: 2) reflect this situation. This finding supports the important role of performance expectancy in technology acceptance as noted by Wu et al. (2022). The second is the expectation of effort. Russian students are more cautious about the use of AI. Ismatullaev and Kim (2024)‘s study similarly emphasizes that effort expectancy is an important factor in technology acceptance. The fact that the “lack of knowledge” code is higher among Russian students (Russia: 14, Turkey: 8) shows this situation. In studies, lack of knowledge is seen as an important barrier in the use of AI (Acosta-Enriquez et al., 2024).

The different viewpoints and life experiences of Turkish and Russian students can be analyzed by considering cultural factors, such as uncertainty avoidance, power distance, and individualism–collectivism (Hofstede, 2001). For instance, the increased focus on privacy concerns among Russian students may be associated with high levels of uncertainty avoidance in Russian culture (Ismatullaev and Kim, 2024). Turkish culture’s collectivist tendencies contribute to Turkish students’ positive views towards cooperation and participation (Keller, 2009). Nations and cultures perceive AI differently (Kose and Pavaloiu, 2018).

Hofstede’s theory of cultural dimensions reveals significant differences in two main dimensions. The first one is the Individualism–Collectivism dimension. Turkish students’ tendency to use AI more in group work (Group work—Turkey: 8, Russia: 2) reflects the characteristics of a collectivist culture. This finding coincides with the findings of Na et al. (2023) on the role of cultural values in technology acceptance. The second dimension is uncertainty avoidance. Russian students’ high concerns about data security and privacy (“privacy and data concerns”—Russia: 4, Turkey: 2) indicate a high uncertainty avoidance tendency. This finding is in line with Grassini and Ree (2023)‘s study on the role of cultural differences in the perception of AI.

The results of this study have significant ramifications for the advancement and execution of AI technologies in higher education. When developing AI-based educational interventions and policies, universities in Turkey, Russia, and other nations should consider the cultural environment. This may require customizing AI systems to meet various learning styles, communication preferences, and ethical issues across diverse cultures. Foltynek et al. (2023) and Yang and Xu (2024) say that schools need to make clear rules and training programs to encourage the responsible and culturally aware use of AI in school settings.

Cultural differences in education systems affect the use of AI. In this context, the differentiation in students’ perceptions of academic integrity supports the hypothesis. While Turkish students address academic integrity from a broader perspective (Academic integrity issues—Turkey: 39, Russia: 30), Russian students express more specific concerns. This disparity could result from the two universities’ different educational systems reflecting different cultural values. This result is consistent with the 2024 research of Yusuf et al. on the influence of cultural viewpoints on the view of academic honesty. An intriguing illustration is the focus on group learning. The collaborative character of the educational system is reflected in the inclination of Turkish students to view AI as a tool for collaborative learning (high values in “group work” and “positive communication”). “This result validates the results of Ho et al. (2023) on the part cultural values play in technology integration.

In essence, both cultural elements and models of technology acceptability intimately connect with the application of AI in academic communication. Our results imply that one cannot regard technological acceptance and use as free from cultural background. Therefore, we should give great attention to cultural elements when formulating regulations for the application of AI in education (see Figure 3).

Figure 3
www.frontiersin.org

Figure 3. Cultural effect model.

Our research results are shown in a three-layered structure in our model, which collects the factors that affect people’s thoughts and feelings about using AI in academic communication. This model reveals that the application of AI is not just a technology problem but also a social and institutional cultural one.

The outer layer of the model includes the three main cultural factors that stand out in our research: power distance, individualism–collectivism, and uncertainty avoidance. Our findings show that these cultural differences between Turkish and Russian students significantly affect the use of AI. For instance, whereas Russian students used AI more as a personal learning tool, Turkish students’ collectivistic perspective drove them to use it more in group projects. Moreover, Russian students’ strong habit of high uncertainty avoidance made them more careful in applying AI.

In the middle part of the model, institutional culture is looked at under three main categories: attitude toward learning, academic integrity, and technology acceptability. Variations in their educational systems directly influence how both nations apply AI Particularly, varied opinions on academic integrity and the way the institutional culture approaches technology seem to be crucial determinants of students’ AI adaptability.

This model is based on the main findings of our research. The model makes it clear that the use of AI in academic communication cannot be considered independent of the cultural and institutional context. The differences in approaches in Turkey and Russia clearly demonstrate the determining role of cultural factors in AI adaptation.

Policy recommendations for the use of artificial intelligence

Our study results imply that colleges should create fresh policies for AI applications. These rules need to support academic honesty as well as inspire creativity. We present concrete policy proposals below, grounded on our research results.

To begin addressing these challenges, universities should establish concrete guidelines for AI use that directly respond to the need for actionable and practical steps. These guidelines must clearly define acceptable and unacceptable uses of AI in different academic contexts, specify which AI tools are permitted for each course, and provide examples. They should integrate AI-related provisions into existing academic integrity policies and actively involve faculty members and students in the policy development process through surveys, focus groups, and workshops.

Education and awareness should be cornerstones of AI policies because they equip students and faculty with the knowledge to navigate AI responsibly. For instance, AI literacy training can help students identify ethical concerns, such as bias in algorithms, while workshops for educators can ensure effective integration of AI tools into teaching practices. Universities should develop and mandate AI literacy training for all students, ensuring they understand the ethical, technical, and practical aspects of AI. Regular faculty workshops should be organized to update educators on emerging AI technologies and best practices for integrating AI into teaching. Seminars and panel discussions on academic integrity and the ethical implications of AI should be hosted, featuring experts from various disciplines. Additionally, case studies of successful AI applications in academia should be shared to inspire innovative approaches.

To ensure accountability, universities should define transparent criteria for evaluating AI-supported academic work, such as assignments and projects. Students should be required to disclose AI usage in their submissions, specifying the tools used and their contributions. Advanced technological tools should be used to verify the authenticity and originality of AI-assisted work, and students and faculty should receive detailed, constructive feedback on the ethical and effective use of AI.

Cultural differences must be considered in AI policy development. For example, some cultures may emphasize collective learning over individual achievement, which could impact how AI tools are integrated into collaborative assignments. Similarly, perceptions of plagiarism and intellectual property may vary, requiring culturally sensitive approaches to academic integrity policies. Institutions should tailor AI guidelines to align with their unique educational traditions and values. Inclusive policies that address diverse learning styles and educational backgrounds should be created, along with ethical standards that respect and incorporate cultural values. Specialized support programs for international students, including orientation sessions on local academic norms and AI usage, should also be developed.

Policies should strike a balance between fostering innovation and maintaining academic integrity. This can be achieved by implementing safeguards such as plagiarism detection tools and clear usage disclosures, alongside incentives like recognition for innovative AI applications in coursework. Encouraging a culture of ethical creativity ensures both goals are met effectively. Universities should encourage creative and innovative uses of AI in research and education while implementing robust safeguards to prevent misuse, such as plagiarism and unauthorized assistance. Habits and practices that enhance student success, such as collaborative learning and critical thinking, should be promoted, ensuring that research and educational activities involving AI adhere to ethical standards.

Universities can refine their AI policies by examining successful examples of AI integration from other institutions, courses, and departments, and adapting these examples to their own institutional needs and cultural contexts. Regularly reviewing and updating AI policies is essential, ideally on an annual basis or at the end of each academic term, to ensure they remain effective and responsive to emerging trends and challenges. Institutions should establish mechanisms for ongoing feedback from faculty and students, periodically assess the impact of AI on learning outcomes and academic integrity, and identify and disseminate effective practices and strategies for AI use.

By adopting these specific and actionable measures, colleges can create policies that effectively manage AI applications, promote academic honesty, and inspire innovation.

Conclusion

This comparison of the opinions of Turkish and Russian university students about AI in academic communication reveals the complex relationship between technology, culture, and knowledge. The results highlight the need to consider cultural features while examining the acceptance and effects of AI in higher education. As AI technologies have a bigger impact on the future of communication and learning, schools need to begin using culturally responsive strategies to make the most of AI’s benefits while also addressing the unique problems and worries of students from different backgrounds. By adopting this approach, institutions may provide inclusive and cutting-edge learning environments that equip students with the necessary skills to thrive in a rapidly evolving world influenced by AI.

Limitations

This study offers insightful analysis, yet there are various factors to take into account even with that. First, given the sample size of 71 university students—33 Turkish and 38 Russian—the full student population in both nations might not be exactly represented. Future studies might want to use a larger and more varied sample to improve the generalizability of the conclusions. Second, the study depended on qualitative data gathered by means of questions, so social desirability bias could have a role. Participants might have answered what they thought were expected or socially acceptable instead of their own views. Future research might use a mixed-methods approach integrating qualitative and quantitative data to minimize this bias. Third, the study concentrated on, at a given moment, university students’ perceptions and experiences. Students’ attitudes and behaviors could shift as AI technology develops and get more included in learning environments. Examining the long-term effects of AI on academic communication and learning results calls for longitudinal research. At last, the study looked at only Turkey and Russia, two nations that would restrict the relevance of the results to other cultural settings. Future studies should cover other nations and educational environments, therefore broadening their topic of inquiry. Also, the proportion of females in the sample is higher than that of males. Researchers should be careful when reporting research results.

Recommendations

The study’s findings suggest several recommendations for future research and practice in the field of AI in higher education. Educational institutions in Turkey, Russia, and other countries can develop AI integration policies taking into account the unique perspectives, concerns, and communication preferences of students from different backgrounds. Universities should promote responsible and ethical use of AI tools in academic settings. They should prioritize the development of clear guidelines and training programs to address issues such as academic integrity, plagiarism, and data privacy.

Future research should further explore the cultural dimensions of AI adoption in higher education. They could examine a wider range of countries and educational contexts to gain a more comprehensive understanding of the phenomenon. It could examine how students’ perceptions and experiences of AI evolve over time. Longitudinal studies could also be conducted to investigate the long-term impact of AI on academic communication and learning outcomes.

Universities should develop clear and understandable policies on the use of AI. These policies should clearly specify which AI tools can be used in which courses and for which purposes. Each university should organize AI literacy trainings for students and faculty. These trainings should address academic honesty, plagiarism, and the ethical application of AI. At the start of the semester, faculty members should let students know their own guidelines for including AI into their classes. Universities ought to create criteria for assessing projects backed by AI. Policies should be developed in line with each nation’s particular educational traditions, considering cultural variances. For instance, societies that value collective projects can promote the application of AI to enhance cooperative learning. In societies that value human effort, rules can be prepared for the use of AI as a personal learning tool. Universities should also routinely evaluate how AI affects learning outcomes and student performance and change their policies in line with these results.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by the Scientific and Educational Center “Person in Communication” of the Pyatigorsk State University. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

OS: Data curation, Formal analysis, Investigation, Methodology, Project administration, Resources, Validation, Visualization, Writing – original draft, Writing – review & editing. MZ: Conceptualization, Data curation, Formal analysis, Methodology, Project administration, Resources, Visualization, Writing – original draft, Writing – review & editing. SD: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Supervision, Visualization, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Acknowledgments

We used Claude 3.5 AI technologies to refine the language. After using this AI technology, we evaluated and validated the final version of our work.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abdelmagid, A. S., Jabli, N. M., and Qahmash, A. I. (2024). The effect of digital platforms and artificial intelligence on the development of engagement skills in learning and DigitalTrust among university students. Am. J. Educ. Technol. 3, 60–69. doi: 10.54536/ajet.v3i1.2459

Crossref Full Text | Google Scholar

Acosta-Enriquez, B. G., Ramos Farroñan, E. V., Villena Zapata, L. I., Mogollon Garcia, F. S., Rabanal-León, H. C., Angaspilco, J. E. M., et al. (2024). Acceptance of artificial intelligence in university contexts: a conceptual analysis based on UTAUT2 theory. Heliyon 10:e38315. doi: 10.1016/j.heliyon.2024.e38315

PubMed Abstract | Crossref Full Text | Google Scholar

Ahmad, S. F., Rahmat Mohd, K., Mubarik, M. S., Alam, M. M., and Hyder, S. (2021). Artificial intelligence and its role in education. Sustain. For. 13:12902. doi: 10.3390/su132212902

Crossref Full Text | Google Scholar

Aldosari, S. A. M. (2020). The future of higher education in the light of artificial intelligence transformations. Int. J. High. Educ. 9, 145–151. doi: 10.5430/ijhe.v9n3p145

Crossref Full Text | Google Scholar

Ali, M. Y., Naeem, S. B., Bhatti, R., and Richardson, J. (2024). Artificial intelligence application in university libraries of Pakistan: SWOT analysis and implications. Glob. Know. Mem. Commun. 73, 219–234. doi: 10.1108/GKMC-12-2021-0203

PubMed Abstract | Crossref Full Text | Google Scholar

Alrayes, A., Henari, T. F., and Ahmed, D. A. (2024). ChatGPT in education – understanding the Bahraini academics perspective. Electr. J. E Learn. 22, 112–134. doi: 10.34190/EJEL.22.2.3250

Crossref Full Text | Google Scholar

American Library Association. (2006). Principles and strategies for the reform of scholarly communication. Available online at: https://www.ala.org/acrl/publications/whitepapers/principlesstrategies (Accessed May 15, 2024).

Google Scholar

Barros, A., Prasad, A., and Śliwa, M. (2023). Generative artificial intelligence and academia: implication for research, teaching and service. Manag. Learn. 54, 597–604. doi: 10.1177/13505076231201445

Crossref Full Text | Google Scholar

Bell, S. (2023). The write algorithm: promoting responsible artificial intelligence usage and accountability in academic writing. BMC Med. 21:334. doi: 10.1186/s12916-023-03039-7

Crossref Full Text | Google Scholar

Bokhari, S. A. A., and Myeong, S. (2023). AI applications in Smart City employing technology adoption model: Hofstede’s cultural perspective. 2023 2nd international conference for innovation in technology (INOCON), 1–6.

Google Scholar

Braun, V., and Clarke, V. (2006). Using thematic analysis in psychology. Qual. Res. Psychol. 3, 77–101. doi: 10.1191/1478088706qp063oa

Crossref Full Text | Google Scholar

Cahyanto, I. (2023). Privacy challenges in using wearable Technology in Education Literature Review. Formosa J. Appl. Sci. 2, 909–928. doi: 10.55927/fjas.v2i6.4272

Crossref Full Text | Google Scholar

Chen, Y., Jensen, S., Albert, L. J., Gupta, S., and Lee, T. (2023). Artificial intelligence (AI) student assistants in the classroom: designing Chatbots to support student success. Inf. Syst. Front. 25, 161–182. doi: 10.1007/s10796-022-10291-4

Crossref Full Text | Google Scholar

Clos, J., and Chen, Y. Y. (2024). Investigating the impact of generative AI on students and educators: evidence and insights from the literature. ACM International Conference Proceeding Series.

Google Scholar

Cotton, D. R. E., Cotton, P. A., and Shipway, J. R. (2024). Chatting and cheating: ensuring academic integrity in the era of ChatGPT. Innov. Educ. Teach. Int. 61, 228–239. doi: 10.1080/14703297.2023.2190148

Crossref Full Text | Google Scholar

Cox, A. M. (2021). Exploring the impact of artificial intelligence and robots on higher education through literature-based design fictions. International journal of educational technology. High. Educ. 18:27. doi: 10.1186/s41239-020-00237-8

Crossref Full Text | Google Scholar

Cox, A. (2023). How artificial intelligence might change academic library work: applying the competencies literature and the theory of the professions. J. Assoc. Inf. Sci. Technol. 74, 367–380. doi: 10.1002/asi.24635

Crossref Full Text | Google Scholar

Creswell, J. W., and Creswell, J. D. (2017). Research design: Qualitative, quantitative, and mixed methods approaches. Cham: Sage publications.

Google Scholar

Curtis, G. J., Gouldthorp, B., Thomas, E. F., O’Brien, G. M., and Correia, H. M. (2013). Online academic-integrity mastery training may improve students’ awareness of, and attitudes toward, plagiarism. Psychol. Learn. Teach. 12, 282–289. doi: 10.2304/plat.2013.12.3.282

Crossref Full Text | Google Scholar

Curtis, G. J., and Vardanega, L. (2016). Is plagiarism changing over time? A 10-year time-lag study with three points of measurement. High. Educ. Res. Dev. 35, 1167–1179. doi: 10.1080/07294360.2016.1161602

Crossref Full Text | Google Scholar

Dai, Y., Liu, A., and Lim, C. P. (2023). Reconceptualizing ChatGPT and generative AI as a student-driven innovation in higher education. Proc. CIRP 119, 84–90. doi: 10.1016/j.procir.2023.05.002

Crossref Full Text | Google Scholar

Darvishi, A., Khosravi, H., Sadiq, S., and Gašević, D. (2022). Incorporating AI and learning analytics to build trustworthy peer assessment systems. Br. J. Educ. Technol. 53, 844–875. doi: 10.1111/bjet.13233

Crossref Full Text | Google Scholar

Dergaa, I., Chamari, K., Zmijewski, P., and Saad, H. B. (2023). From human writing to artificial intelligence generated text: examining the prospects and potential threats of ChatGPT in academic writing. Biol. Sport 40, 615–622. doi: 10.5114/BIOLSPORT.2023.125623

PubMed Abstract | Crossref Full Text | Google Scholar

Devi, S., Angita, D., Boruah, S., Nirban, S., Dushyant Nimavat, D., and Bajaj, M. K. K. (2023). Ethical considerations in using artificial intelligence to improve teaching and learning. J. Propuls. Technol. 44:966. doi: 10.52783/tjjpt.v44.i4.966

Crossref Full Text | Google Scholar

Dwivedi, Y. K., Hughes, L., Ismagilova, E., Aarts, G., Coombs, C., Crick, T., et al. (2021). Artificial intelligence (AI): multidisciplinary perspectives on emerging challenges, opportunities, and agenda for research, practice and policy. Int. J. Inf. Manag. 57:101994. doi: 10.1016/j.ijinfomgt.2019.08.002

Crossref Full Text | Google Scholar

Eden, C. A., Chisom, O. N., and Adeniyi, I. S. (2024). Integrating AI in education: opportunities, challenges, and ethical considerations. Magna Sci. Adv. Res. Rev. 10, 006–013. doi: 10.30574/msarr.2024.10.2.0039

Crossref Full Text | Google Scholar

Foltynek, T., Bjelobaba, S., Glendinning, I., Khan, Z. R., Santos, R., Pavletic, P., et al. (2023). ENAI recommendations on the ethical use of artificial intelligence in education. Int. J. Educ. Integrity 19:33. doi: 10.1007/s40979-023-00133-4

Crossref Full Text | Google Scholar

Fowler, D. S. (2023). AI in higher education. J. Ethics High. Educ. 3, 127–143. doi: 10.26034/fr.jehe.2023.4657

PubMed Abstract | Crossref Full Text | Google Scholar

Geethalakshmi, S. (2018). Dealing with academic dishonesty. Bioethics Enq. 6:2. doi: 10.38020/GBE.6.2.2018.71-75

Crossref Full Text | Google Scholar

Goel, R. K., and Nelson, M. A. (2024). Do college anti-plagiarism/cheating policies have teeth in the age of AI? Exploratory evidence from the internet. Manag. Decis. Econ. 45, 2336–2347. doi: 10.1002/mde.4139

Crossref Full Text | Google Scholar

Grassini, S., and Ree, A. S. (2023). Hope or doom AI-ttitude? Examining the impact of gender, age, and cultural differences on the envisioned future impact of artificial intelligence on humankind. ACM International Conference Proceeding Series.

Google Scholar

Guan, X., Feng, X., and Islam, A. Y. M. A. (2023). The dilemma and countermeasures of educational data ethics in the age of intelligence. Humanities and social sciences. Communications 10:138. doi: 10.1057/s41599-023-01633-x

PubMed Abstract | Crossref Full Text | Google Scholar

Hannan, E., and Liu, S. (2023). AI: new source of competitiveness in higher education. Compet. Rev. 33, 265–279. doi: 10.1108/CR-03-2021-0045

Crossref Full Text | Google Scholar

Harry, A. (2023). Role of AI in education. Interdicipl. J. Humm. 2, 260–268. doi: 10.58631/injurity.v2i3.52

Crossref Full Text | Google Scholar

Ho, M. T., Le, N. T. B., Mantello, P., Ho, M. T., and Ghotbi, N. (2023). Understanding the acceptance of emotional artificial intelligence in Japanese healthcare system: a cross-sectional survey of clinic visitors’ attitude. Technol. Soc. 72:102166. doi: 10.1016/j.techsoc.2022.102166

Crossref Full Text | Google Scholar

Hofstede, G. (2001). Culture’s consequences: Comparing values, behaviors, institutions and organizations across nations. Cham: Sage publications.

Google Scholar

Hohenstein, J., Kizilcec, R. F., DiFranzo, D., Aghajari, Z., Mieczkowski, H., Levy, K., et al. (2023). Artificial intelligence in communication impacts language and social relationships. Sci. Rep. 13:5487. doi: 10.1038/s41598-023-30938-9

PubMed Abstract | Crossref Full Text | Google Scholar

Huang, L. (2023). Ethics of artificial intelligence in education: student privacy and data protection. Sci. Insights Educ. Front. 16, 2577–2587. doi: 10.15354/sief.23.re202

Crossref Full Text | Google Scholar

Huang, J., and Tan, M. (2023). The role of ChatGPT in scientific communication: writing better scientific review articles. Am. J. Cancer Res. 13, 1148–1154.

PubMed Abstract | Google Scholar

Iffath Unnisa Begum (2024). Role of artificial intelligence in higher education- an empirical investigation. Int. Res. J. Adv. Eng. Manag. 2, 49–53. doi: 10.47392/IRJAEM.2024.0009

PubMed Abstract | Crossref Full Text | Google Scholar

Ismatullaev, U. V. U., and Kim, S. H. (2024). Review of the factors affecting acceptance of AI-infused systems. Hum. Factors 66, 126–144. doi: 10.1177/00187208211064707

PubMed Abstract | Crossref Full Text | Google Scholar

Jaboob, M., Hazaimeh, M., and Al-Ansi, A. M. (2024). Integration of generative AI techniques and applications in student behavior and cognitive achievement in Arab higher education. Int. J. Hum. Comput. Int. 41, 353–366. doi: 10.1080/10447318.2023.2300016

PubMed Abstract | Crossref Full Text | Google Scholar

Jain, R., Garg, N., and Khera, S. N. (2022). Adoption of AI-enabled tools in social development organizations in India: an extension of UTAUT model. Front. Psychol. 13:893691. doi: 10.3389/fpsyg.2022.893691

PubMed Abstract | Crossref Full Text | Google Scholar

Jia, X. H., and Tu, J. C. (2024). Towards a new conceptual model of AI-enhanced learning for college students: the roles of artificial intelligence capabilities, general self-efficacy, learning motivation, and critical thinking awareness. Systems 12:74. doi: 10.3390/systems12030074

Crossref Full Text | Google Scholar

Kacena, M. A., Plotkin, L. I., and Fehrenbacher, J. C. (2024). The use of artificial intelligence in writing scientific review articles. Curr. Osteop. Rep. 22, 115–121. doi: 10.1007/s11914-023-00852-0

PubMed Abstract | Crossref Full Text | Google Scholar

Karkoulian, S., Sayegh, N., and Sayegh, N. (2024). ChatGPT unveiled: understanding perceptions of academic integrity in higher education - a qualitative approach. J. Acad. Ethics. doi: 10.1007/s10805-024-09543-6

Crossref Full Text | Google Scholar

Keller, C. (2009). User acceptance of virtual learning environments: a case study from three northern European universities. Commun. Assoc. Inf. Syst. 25, 465–486. doi: 10.17705/1cais.02538

Crossref Full Text | Google Scholar

Khalil, M., and Er, E. (2023). “Will ChatGPT Get You Caught? Rethinking of Plagiarism Detection,” in Learning and Collaboration Technologies. HCII 2023. Lecture Notes in Computer Science, vol 14040. eds. P. Zaphiris and A. Ioannou (Cham: Springer).

Google Scholar

Khan, M. A., Khojah, M., and Vivek, A. (2022). Artificial intelligence and big data: the advent of new pedagogy in the adaptive E-learning system in the higher educational institutions of Saudi Arabia. Educ. Res. Int. 2022, 1–10. doi: 10.1155/2022/1263555

PubMed Abstract | Crossref Full Text | Google Scholar

Kim, J., and Lee, S. S. (2023). Are two heads better than one?: the effect of student-AI collaboration on students’ learning task performance. TechTrends 67, 365–375. doi: 10.1007/s11528-022-00788-9

PubMed Abstract | Crossref Full Text | Google Scholar

Kose, U., and Pavaloiu, A. (2018). A cross-cultural perspective on the societal impact of artificial intelligence. Proceedings of the 6th International Virtual Conference on Advanced Scientific Results.

Google Scholar

Krishnamoorthy, S., Tr, E., Muruganathan, A., Ramakrishan, S., Nanda, S., and Radhakrishnan, P. (2022). The impact of cultural dimensions of clinicians on the adoption of artificial intelligence in healthcare. J. Assoc. Physicians India 70, 11–12

PubMed Abstract | Google Scholar

Li, B., Bonk, C. J., Wang, C., and Kou, X. (2024). Reconceptualizing self-directed learning in the era of generative AI: an exploratory analysis of language learning. IEEE Trans. Learn. Technol. 17, 1489–1503. doi: 10.1109/TLT.2024.3386098

PubMed Abstract | Crossref Full Text | Google Scholar

Li, Y., Wu, B., Huang, Y., and Luan, S. (2024). Developing trustworthy artificial intelligence: insights from research on interpersonal, human-automation, and human-AI trust. Front. Psychol. 15:1382693. doi: 10.3389/fpsyg.2024.1382693

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, Y. (2023). Cross-cultural challenges to artificial intelligence ethics. Comput. Sci. Math. Forum 8:21. doi: 10.3390/cmsf2023008021

Crossref Full Text | Google Scholar

Lund, B. D., and Naheem, K. T. (2024). Can ChatGPT be an author? A study of artificial intelligence authorship policies in top academic journals. Learn. Publish. 37, 13–21. doi: 10.1002/leap.1582

Crossref Full Text | Google Scholar

Lund, B. D., Wang, T., Mannuru, N. R., Nie, B., Shimray, S., and Wang, Z. (2023). ChatGPT and a new academic reality: artificial intelligence-written research papers and the ethics of the large language models in scholarly publishing. J. Assoc. Inf. Sci. Technol. 74, 570–581. doi: 10.1002/asi.24750

Crossref Full Text | Google Scholar

Ma, X., and Jiang, C. (2023). On the ethical risks of artificial intelligence applications in education and its avoidance strategies. J. Educ. Hum. Soc. Sci. SSMI 14, 354–359. doi: 10.54097/ehss.v14i.8868

Crossref Full Text | Google Scholar

Mansoor, H. M. H., Bawazir, A., Alsabri, M. A., Alharbi, A., and Okela, A. H. (2024). Artificial intelligence literacy among university students—a comparative transnational survey. Front. Commun. 9:1478476. doi: 10.3389/fcomm.2024.1478476

Crossref Full Text | Google Scholar

Meszaros, J., and Ho, C. (2021). AI research and data protection: can the same rules apply for commercial and academic research under the GDPR? Comput. Law Secur. Rev. 41:105532:105532. doi: 10.1016/j.clsr.2021.105532

Crossref Full Text | Google Scholar

Miles, M. B., and Huberman, A. M. (1994). Qualitative data analysis: An expanded sourcebook. Sage.

Google Scholar

Miao, J., Thongprayoon, C., Suppadungsuk, S., Garcia Valencia, O. A., Qureshi, F., and Cheungpasitporn, W. (2024). Ethical dilemmas in using AI for academic writing and an example framework for peer review in nephrology academia: a narrative review. Nephrol. Acad. 2023, 14, 89–105. doi: 10.3390/clinpract

Crossref Full Text | Google Scholar

Na, S., Heo, S., Choi, W., Kim, C., and Whang, S. W. (2023). Artificial intelligence (AI)-based technology adoption in the construction industry: a cross National Perspective Using the technology acceptance model. Buildings 13:518. doi: 10.3390/buildings13102518

Crossref Full Text | Google Scholar

Neumann, O., Guirguis, K., and Steiner, R. (2024). Exploring artificial intelligence adoption in public organizations: a comparative case study. Public Manag. Rev. 26, 114–141. doi: 10.1080/14719037.2022.2048685

Crossref Full Text | Google Scholar

Nikolopoulou, K. (2024). Generative artificial intelligence in higher education: exploring ways of harnessing pedagogical practices with the assistance of ChatGPT. Int. J. Changes Educ. 1, 103–111. doi: 10.47852/bonviewijce42022489

Crossref Full Text | Google Scholar

Okonkwo, C. W., and Ade-Ibijola, A. (2021). Chatbots applications in education: a systematic review. Comput. Educ. 2:100033. doi: 10.1016/j.caeai.2021.100033

PubMed Abstract | Crossref Full Text | Google Scholar

Okunlaya, R. O., Syed Abdullah, N., and Alias, R. A. (2022). Artificial intelligence (AI) library services innovative conceptual framework for the digital transformation of university education. Lib. Hi Tech 40, 1869–1892. doi: 10.1108/LHT-07-2021-0242

Crossref Full Text | Google Scholar

Owoc, M. L., Sawicka, A., and Weichbroth, P. (2021). “Artificial intelligence technologies in education: Benefits, challenges and strategies of implementation,” in Artificial intelligence for knowledge management. AI4KM 2019. IFIP advances in information and communication technology, vol 599. eds. M. L. Owoc and M. Pondel (Cham: Springer).

Google Scholar

Passmore, J., and Tee, D. (2024). The library of babel: assessing the powers of artificial intelligence in knowledge synthesis, learning and development and coaching. J. Work Appl. Manag. 16, 4–18. doi: 10.1108/JWAM-06-2023-0057

Crossref Full Text | Google Scholar

Payini, V., Kamath, G. B., Kamath, V., and Mallya, J. (2024). Sense of community and community development: An integrative review. Community Psychol. Glob. Perspect. 10, 177–203.

Google Scholar

Perkins, M. (2023). Academic integrity considerations of AI large language models in the post-pandemic era: ChatGPT and beyond. J. Univ. Teach. Learn. Pract. 20, 1–24. doi: 10.53761/1.20.02.07

Crossref Full Text | Google Scholar

Perkins, M., and Roe, J. (2024). Decoding academic integrity policies: a Corpus linguistics investigation of AI and other technological threats. High Educ. Pol. 37, 633–653. doi: 10.1057/s41307-023-00323-2

Crossref Full Text | Google Scholar

Petroșanu, D. M., Pîrjan, A., and Tăbușcă, A. (2023). Tracing the influence of large language models across the most impactful scientific works. Electronics 12:957. doi: 10.3390/electronics12244957

Crossref Full Text | Google Scholar

Porter, B., and Grippa, F. (2020). A platform for ai-enabled real-time feedback to promote digital collaboration. Sustainability 12, 1–13. doi: 10.3390/su122410243

PubMed Abstract | Crossref Full Text | Google Scholar

Rabbianty, E. N., Azizah, S., and Virdyna, N. K. (2023). AI in academic writing: assessing current usage and future implications. INSANIA 28, 14–35. doi: 10.24090/insania.v28i1a.9278

PubMed Abstract | Crossref Full Text | Google Scholar

Risang Baskara, F. X. (2023). Personalised learning with AI: implications for ignatian pedagogy. Int. J. Educ. Best Pract. 7, 1–16. doi: 10.32851/ijebp.v7n1.p1-16

Crossref Full Text | Google Scholar

Robinson, S. C. (2020). Trust, transparency, and openness: how inclusion of cultural values shapes Nordic national public policy strategies for artificial intelligence (AI). Technol. Soc. 63:101421. doi: 10.1016/j.techsoc.2020.101421

Crossref Full Text | Google Scholar

Rui, Z., and Badarch, T. (2022). Research on applications of artificial intelligence in education. Am. J. Comput. Sci. Technol. 5:72. doi: 10.11648/j.ajcst.20220502.17

Crossref Full Text | Google Scholar

Sarwar, M. A., Saima, M., and Gul, A. (2024). The role of artificial intelligence in shaping the future of education at higher secondary level. J. Educ. Soc. Stud. 5, 34–45. doi: 10.52223/jess.2024.5104

Crossref Full Text | Google Scholar

Sarwari, A. Q., Javed, M. N., Mohd Adnan, H., and Abdul Wahab, M. N. (2024). Assessment of the impacts of artificial intelligence (AI) on intercultural communication among postgraduate students in a multicultural university environment. Sci. Rep. 14:13849. doi: 10.1038/s41598-024-63276-5

PubMed Abstract | Crossref Full Text | Google Scholar

Sergeeva, O. V., Zheltukhina, M. R., Shoustikova, T., Tukhvatullina, L. R., Dobrokhotov, D. A., and Kondrashev, S. V. (2025). Understanding higher education students’ adoption of generative AI technologies: An empirical investigation using UTAUT2. Contemp. Educ. Technol. 17:ep571. doi: 10.30935/cedtech/16039

Crossref Full Text | Google Scholar

Seo, K., Tang, J., Roll, I., Fels, S., and Yoon, D. (2021). The impact of artificial intelligence on learner–instructor interaction in online learning. Int. J. Educ. Technol. High. Educ. 18:54. doi: 10.1186/s41239-021-00292-9

PubMed Abstract | Crossref Full Text | Google Scholar

Shrivastava, R., and Mahajan, P. (2021). Influence of social networking sites on scholarly communication: a study using literature in artificial intelligence. J. Librariansh. Inf. Sci. 53, 522–529. doi: 10.1177/0961000616678309

Crossref Full Text | Google Scholar

Simpson, D. (2023). Educators, students, and plagiarism in age of AI. BMJ 381:p1403. doi: 10.1136/bmj.p1403

PubMed Abstract | Crossref Full Text | Google Scholar

Son, J., and Jin, T. (2024). How to exploit China’s AI-powered platforms for Korean-Chinese translation/interpreting education. INContext 4, 31–52. doi: 10.54754/incontext.v4i1.75

Crossref Full Text | Google Scholar

Song, N. Y. (2024). Higher education crisis: academic misconduct with generative AI. J. Conting. Crisis Manag. 32:532. doi: 10.1111/1468-5973.12532

Crossref Full Text | Google Scholar

Sontan, A. D., and Samuel, S. V. (2024). The intersection of artificial intelligence and cybersecurity: challenges and opportunities. World J. Adv. Res. Rev. 21, 1720–1736. doi: 10.30574/wjarr.2024.21.2.0607

Crossref Full Text | Google Scholar

Stöhr, C., Ou, A. W., and Malmström, H. (2024). Perceptions and usage of AI chatbots among students in higher education across genders, academic levels and fields of study. Comput. Educ. 7:100259. doi: 10.1016/j.caeai.2024.100259

PubMed Abstract | Crossref Full Text | Google Scholar

Sullivan, M., Kelly, A., and McLaughlan, P. (2023). ChatGPT in higher education: considerations for academic integrity and student learning. J. Appl. Learn. Teach. 6, 31–40. doi: 10.37074/jalt.2023.6.1.17

Crossref Full Text | Google Scholar

Tang, K. H. D. (2024). Implications of artificial intelligence for teaching and learning. Acta Pedag. Asia 3, 65–79. doi: 10.53623/apga.v3i2.404

Crossref Full Text | Google Scholar

Thomas, T. (2024). The role of artificial intelligence in formal and informal education for students. Int. J. Res. Appl. Sci. Eng. Technol. 12, 69–71. doi: 10.22214/ijraset.2024.58738

Crossref Full Text | Google Scholar

Tierney, A., Peasey, P., and Gould, J. (2025). Student perceptions on the impact of AI on their teaching and learning experiences in higher education. Res. Pract. Technol. Enhanc. Learn. 20:5. doi: 10.58459/rptel.2025.20005

Crossref Full Text | Google Scholar

Tomar, P., and Verma, S. (2021). “Impact and role of AI technologies in teaching, learning, and research in higher education,” in Impact of AI Technologies on Teaching, Learning, and Research in Higher Education. IGI Global. p. 190–203.

Google Scholar

Venkatesh, V. (2000). Determinants of perceived ease of use: integrating control, intrinsic motivation, and emotion into the technology acceptance model. Inform. Syst. Res. 11, 342–365. doi: 10.1287/isre.11.4.342.11872

Crossref Full Text | Google Scholar

Villegas-Ch, W., García-Ortiz, J., Mullo-Ca, K., Sánchez-Viteri, S., and Roman-Cañizares, M. (2021). Implementation of a virtual assistant for the academic management of a university with the use of artificial intelligence. Future Internet 13:97. doi: 10.3390/fi13040097

Crossref Full Text | Google Scholar

Wang, T., Lund, B. D., Marengo, A., Pagano, A., Mannuru, N. R., Teel, Z. A., et al. (2023). Exploring the potential impact of artificial intelligence (AI) on international students in higher education: generative AI, Chatbots, analytics, and international student success. Appl. Sci. 13:716. doi: 10.3390/app13116716

Crossref Full Text | Google Scholar

Watanabe, A. (2024). Have courage to use your own mind, with or without AI: the relevance of Kant’s enlightenment to higher education in the age of artificial intelligence. Electr. J. E Learn. 22, 46–58. doi: 10.34190/ejel.21.5.3229

PubMed Abstract | Crossref Full Text | Google Scholar

William, K. A. (2024). What is plagiarism in academic writing and how to avoid it: guiding novice researchers. Int. J. Res. Public. 144:161. doi: 10.47119/ijrp1001441320246161

Crossref Full Text | Google Scholar

Wolfram, D. (2019). Understanding and navigating the scholarly communication landscape in the twenty-first century. Front. Res. Metrics Anal. 4:4. doi: 10.3389/frma.2019.00004

PubMed Abstract | Crossref Full Text | Google Scholar

Wu, W., Zhang, B., Li, S., and Liu, H. (2022). Exploring factors of the willingness to accept AI-assisted learning environments: an empirical investigation based on the UTAUT model and perceived risk theory. Front. Psychol. 13:777. doi: 10.3389/fpsyg.2022.870777

PubMed Abstract | Crossref Full Text | Google Scholar

Yam, K. C., Tan, T., Jackson, J. C., Shariff, A., and Gray, K. (2023). Cultural differences in people s reactions and applications of robots, algorithms, and artificial intelligence. Manag. Organ. Rev. 19, 859–875. doi: 10.1017/mor.2023.21

Crossref Full Text | Google Scholar

Yan, D. (2023). Impact of ChatGPT on learners in a L2 writing practicum: an exploratory investigation. Educ. Inf. Technol. 28, 13943–13967. doi: 10.1007/s10639-023-11742-4

Crossref Full Text | Google Scholar

Yan, W., Nakajima, T., and Sawada, R. (2024). Benefits and challenges of collaboration between students and conversational generative artificial intelligence in programming learning: an empirical case study. Educ. Sci. 14:433. doi: 10.3390/educsci14040433

Crossref Full Text | Google Scholar

Yang, Y., and Xu, X. (2024). Research on artificial intelligence literacy level and its influencing factors of high school students. Int. J. Soc. Sci. Res. 2. doi: 10.58531/ijssr/2/2/1

Crossref Full Text | Google Scholar

Ye, T., Xue, J., He, M., Gu, J., Lin, H., Xu, B., et al. (2019). Psychosocial factors affecting artificial intelligence adoption in health care in China: cross-sectional study. J. Med. Internet Res. 21:e14316. doi: 10.2196/14316

PubMed Abstract | Crossref Full Text | Google Scholar

Yin, R. K. (2018). Case study research and applications. Thousand Oaks, CA: Sage.

Google Scholar

Yusuf, A., Pervin, N., and Román-González, M. (2024). Generative AI and the future of higher education: a threat to academic integrity or reformation? Evidence from multicultural perspectives. Int. J. Educ. Technol. High. Educ. 21:21. doi: 10.1186/s41239-024-00453-6

PubMed Abstract | Crossref Full Text | Google Scholar

Zaitsu, W., and Jin, M. (2023). Distinguishing ChatGPT(−3.5, −4)-generated and human-written papers through Japanese stylometric analysis. PLoS One 18:e0288453. doi: 10.1371/journal.pone.0288453

PubMed Abstract | Crossref Full Text | Google Scholar

Zawacki-Richter, O., Marín, V. I., Bond, M., and Gouverneur, F. (2019). Systematic review of research on artificial intelligence applications in higher education – where are the educators? Int. J. Educ. Technol. High. Educ. 16, 1–27. doi: 10.1186/s41239-019-0171-0

Crossref Full Text | Google Scholar

Zhang, C., and Qi, Y. (2024). AIED in English preparatory courses: a comparative study of UK and north American universities and its impact on Chinese international student mobility. Lect. Notes Educ. Psychol. Public Media 54, 65–71. doi: 10.54254/2753-7048/54/20241550

Crossref Full Text | Google Scholar

Zhao, Y., and Xue, Y. (2024). Research on the teaching reform of art and design courses in colleges and universities driven by artificial intelligence. Appl. Math. Nonlinear Sci. 9. doi: 10.2478/amns-2024-0625

PubMed Abstract | Crossref Full Text | Google Scholar

Zheng, L. (2017). Does online perceived risk depend on culture? Individualistic versus collectivistic culture. J. Decis. Syst. 26, 256–274. doi: 10.1080/12460125.2017.1351861

Crossref Full Text | Google Scholar

Keywords: artificial intelligence, academic communication, academic integrity, higher education, cultural studies

Citation: Sergeeva OV, Zheltukhina MR and Demir S (2025) From assistance to integrity: exploring the role of AI in academic communication across Russian and Turkish campuses. Front. Commun. 10:1479813. doi: 10.3389/fcomm.2025.1479813

Received: 12 August 2024; Accepted: 28 April 2025;
Published: 26 May 2025.

Edited by:

Musa Adekunle Ayanwale, University of Johannesburg, South Africa

Reviewed by:

Rui Cruz, European University of Lisbon, Portugal
Maria Joseph Israel, Santa Clara University, United States

Copyright © 2025 Sergeeva, Zheltukhina and Demir. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Servet Demir, c2VydmV0ZGVtaXJ0cjI3QGdtYWlsLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.