- Department of Clinical Laboratory, Hangzhou Third People’s Hospital, Hangzhou, Zhejiang, China
Introduction: Undergraduate clinical medical interns often lack systematic laboratory medicine training, potentially impacting their diagnostic reasoning and patient safety. This study aimed to assess the perceived knowledge and attitudes toward clinical laboratory medicine among this population in China, addressing a significant gap in medical education evaluation.
Methods: A cross-sectional study was conducted from March to April 2025 across 11 general hospitals in Eastern China (Shanghai, Hangzhou, Nanjing, Suzhou, Ningbo, Xuzhou, Shaoxing, Yangzhou, Huzhou, and Taizhou). The self-developed and validated 13-item Clinical Laboratory Knowledge and Attitudes Questionnaire (CLKAQ) was structured in three domains: Knowledge, Attitudes and Suggestions. All 303 clinical interns completed the instrument. SPSS 25.0 and AMOS 26.0 were used. Descriptive statistics (frequencies/percentages for qualitative data; mean ± SD for quantitative data) summarized characteristics, knowledge, and attitudes. Scale reliability and validity were confirmed. Normality was assessed via Kolmogorov–Smirnov test. Group comparisons (gender, age, city tier) employed Mann–Whitney U and Kruskal-Wallis tests. Spearman correlation examined knowledge-attitude relationships. Multiple response and content analysis supplemented quantitative findings.
Results: The mean self-perceived knowledge scale score (5-point Likert scale) among the 303 interns was 2.22 ± 0.424. The mean attitude scale score (5-point Likert scale) was 4.05 ± 0.312. Significant differences emerged in key competencies: Gender disparities in report interpretation (Q3), perceived importance of laboratory knowledge (Q5), and learning motivation (Q7); Age-group variations in perceived knowledge adequacy (Q1), (Q5) and (Q7); Interns from third-tier cities demonstrated consistently higher self-perceived competence across all knowledge and attitude dimensions than those in tier-1/2 cities (p < 0.05). A weak positive correlation linked knowledge and attitude levels (r = 0.171, p < 0.05). Critical differences were noted in preferred learning channels (Q10) and perceived barriers (Q11). Regarding open-ended questions, all interns expressed the need for increased clinical laboratory knowledge training and provided specific suggestions for such training.
Conclusion: Undergraduate clinical interns demonstrated suboptimal clinical laboratory knowledge but expressed highly positive attitudes toward learning. This underscores the critical need to enhance clinical laboratory training during clerkship. Implementing measures to improve knowledge is necessary. These findings inform curriculum optimization and educational strategy development for clinical continuing education.
1 Introduction
The Accreditation Council for Graduate Medical Education (ACGME) requires residents to demonstrate knowledge of established and evolving biomedical, clinical, epidemiological, and social-behavioral sciences—encompassing scientific inquiry—and apply this knowledge to patient care (1). Clinical residency education typically progresses through three stages: undergraduate medical education, clerkship, and standardized residency training. However, persistent challenges exist—undergraduate education (2), clerkship (3), and standardized training (4–6)—such as the need to improve interns’ diagnostic and therapeutic capabilities, and their ability to integrate and analyze laboratory results in the diagnostic process. A fundamental requirement for graduates is the ability to select appropriate clinical laboratory tests based on patient condition, safety, and cost-effectiveness, justify their selection, and interpret the results (7). Clinical laboratory testing is central to disease diagnosis and treatment. Within the current medical education system, clinical medical students receive minimal medical laboratory science coursework or lack systematic rotation training within the clinical laboratory department, resulting in insufficient laboratory knowledge. The assessment of clinical reasoning and medical knowledge is becoming increasingly crucial in clinical practice, primarily due to the persistent and harmful problem of diagnostic errors (8, 9). Global studies confirm systemic issues: Saffar et al. (10) reported 47.9% of Iranian students self-rated as ‘weak’ in test interpretation; Alosaimi (11) found only 18.3% of Saudi interns passed basic microbiology assessments; Twisha et al. (12) documented 15% inappropriate test utilization in Australian hospitals. Despite these challenges and laboratory medicine’s critical role (9, 13) no study has comprehensively assessed self-perceived knowledge-attitude discrepancies among Chinese clinical interns.
To address this gap, we conducted the Clinical Laboratory Knowledge and Attitudes Questionnaire (CLKAQ) instrument with three key advancements over prior research: First multidimensional assessment of knowledge and attitudes among Chinese clinical interns; Samples come from different city tiers and direct prioritization of training needs from 303 interns’ quantitative/qualitative feedback. This study aims to analyze undergraduate clinical medicine interns’ knowledge and attitudes regarding clinical laboratory medicine through a questionnaire survey. The findings will highlight the necessity of enhancing laboratory medicine education within undergraduate clinical training and provide a reference for further optimizing interns’ continuing education curricula, educational improvements, and strategy development.
2 Methods
2.1 Questionnaire design and validation
Data were collected using a cross-sectional survey supplemented with qualitative content analysis of open-ended responses. Surveys as tools are widely used in Health Professions’ Education (HPE) (14). The Clinical Laboratory Knowledge and Attitudes Questionnaire (CLKAQ), specifically developed and validated for this study targeting undergraduate clinical medicine interns (Questionnaire details are provided in Supplementary Material 1), was constructed based on: the Assessment Guidebook (8) detailing the ACGME use of milestones for resident assessment; the ACGME Common Program Requirements (Residency) (1); the overall objectives outlined in the Chinese Undergraduate Medical Education Standards—Standards for basic medical education in China (2022 Edition) (7); a review of existing literature (10, 15, 16); and findings derived from focus group discussions and pilot testing.
The questionnaire comprised four sections: Part I collected participants’ basic information; Part II assessed self-perceived clinical laboratory knowledge through four items (Q1-Q4) utilizing a 5-point Likert scale (“Strongly Disagree” to “Strongly Agree”), evaluating perceived adequacy of knowledge for clinical job requirements, clarity of position-specific expectations, ability to independently interpret common reports (e.g., CBC, biochemistry, coagulation), and implementation capability of the clinical laboratory critical value reporting system; Part III measured attitudes via three Likert-scale items (Q5-Q7) addressing perceived importance of laboratory knowledge, with four additional non-scale multiple-choice items (Q8-Q11) investigating preferred knowledge domains, learning motivations, knowledge acquisition channels, and barriers. The final section (Part IV) consisted of two open-ended items: one eliciting a binary response regarding educational enhancement needs, followed by an item requesting specific training suggestions.
A multidisciplinary expert panel was convened, comprising 10 specialists: 2 medical educators, 2 residency program directors, 4 clinical specialists, and 2 laboratory medicine specialists. A pilot test was conducted with 31 clinical interns (excluded from the main study), confirming item clarity and reliability (Cronbach’s α > 0.7 reliability threshold) (17, 18).
The CLKAQ was developed and administered electronically via the WeChat-based Questionnaire Star mini-program. A QR code linked to the survey was generated within the platform and disseminated simultaneously to all potential participants. Participants accessed the questionnaire by scanning the QR code using WeChat. This process yielded 324 submitted responses. After data screening excluding 21 invalid entries, 303 valid questionnaires were retained for analysis (dropout rate of 6.48%). The CLKAQ’s psychometric properties were assessed using data from all 303 participants, yielding the following metrics for reliability and validity.
The Cronbach’s α for the knowledge dimension was 0.905 and for the attitudes dimension was 0.803, both exceeding the conventional threshold of 0.7 for acceptable internal consistency (17, 18); The Intraclass Correlation Coefficients (ICC) indicated moderate reliability for both the knowledge (ICC = 0.705) and attitudes (ICC = 0.576) dimensions, based on conventional thresholds (19). These thresholds define reliability as poor (<0.5), moderate (0.5–0.75), good (0.75–0.90), or excellent (>0.90). The 95% confidence intervals for these estimates are provided in Table 1.
Content validity was assessed by a panel of 10 experts. For each item, the Content Validity Ratio (CVR) was computed. Based on the recalculated critical values by Wilson et al. (20)(which corrected anomalies in the original Lawshe’s table), the critical CVR value for a one-tailed test at α = 0.05 was 0.520. The Item-Content Validity Index (I-CVI) was also calculated for each item, for which a value of ≥0.78 was deemed acceptable according to established guidelines (21). As detailed in Supplementary Material 2, all items demonstrated excellent content validity: all I-CVI values ranged from 0.90 to 1.00, and all CVR values exceeded the critical threshold of 0.520 (range: 0.80 to 1.00). Finally, the Scale-Content Validity Index/Average (S-CVI/Ave) was 0.97, indicating excellent content validity at the scale level (21).
Exploratory Factor Analysis (EFA) was conducted to examine the underlying factor structure of the questionnaire. The suitability of the data for EFA was confirmed by a Kaiser-Meyer-Olkin (KMO) measure of 0.783, exceeding the recommended threshold of 0.70 for factorability (22), and a statistically significant Bartlett’s test of sphericity (χ2 = 1574.626, p < 0.001) (23). Principal Axis Factoring was selected as the extraction method, consistent with the goal of identifying latent constructs. The analysis yielded a clear two-factor solution, which cumulatively explained 79.4% of the total variance. All items demonstrated strong and salient loadings on their primary factors (range: 0.76 to 0.92; Table 1), with no significant cross-loadings (all<0.30). The pattern of loadings cleanly corresponded to the pre-defined Knowledge and Attitudes dimensions, resulting in a simple structure with factors that are statistically independent.
2.2 Dimension correlation analysis
Spearman’s correlation analysis revealed a statistically significant weak positive correlation between the knowledge and attitudes dimensions (r = 0.171, p < 0.05; Table 2).
CFA analysis in AMOS 26.0 validated the questionnaire’s factor structure, with post-modification fit indices for all six established criteria: the chi-square to degrees of freedom ratio (χ2/df), Root Mean Square Error of Approximation (RMSEA), Goodness of Fit Index (GFI), Normed Fit Index (NFI), Comparative Fit Index (CFI) and Tucker-Lewis Index (TLI) demonstrating satisfactory two-factor convergent validity and model fit, as detailed in Table 3. The confirmatory factor analysis demonstrated an acceptable fit of the two-factor model to the data. Although the RMSEA value of 0.063 slightly exceeded the stringent cutoff of 0.06 proposed by Hu and Bentler (24), it is well within the ‘reasonable fit’ range (0.05–0.08) as recommended by MacCallum et al. (25). This, combined with the excellent performance of other key indices (CFI = 0.995, >0.95; TLI = 0.984, > 0.95; χ2/df = 2.214, < 3), provides strong evidence for the construct validity of the proposed two-factor structure.
2.3 Sample
This cross-sectional online study assessed clinical laboratory knowledge and attitudes among undergraduate clinical medicine interns in China. Participants were recruited between March 28 and April 28, 2025, during the mid-to-late clerkship phase when interns had completed rotations across most hospital departments. The self-administered online questionnaire (CLKAQ) was distributed via the Questionnaire Star platform as a single cross-sectional survey. Eleven general hospitals representing varying economic development tiers (tier-1: Shanghai, Hangzhou, Nanjing; tier-2: Suzhou, Ningbo, Xuzhou, Shaoxing; tier-3: Yangzhou, Huzhou, Taizhou) were randomly selected. The study population comprised interns from the following hospitals: Shanghai Yangpu District Hospital, Nanjing Jiangning Hospital, The First Affiliated Hospital of Zhejiang University, Hangzhou Third People’s Hospital, Suzhou Wujiang District First Hospital, Ningbo Yinzhou District Hospital, Xuzhou Mining Group General Hospital, Shaoxing People’s Hospital, Yangzhou First People’s Hospital, Huzhou Central Hospital, Taizhou Central Hospital.
Inclusion Criteria: (1) Age ≥20 years; (2) Voluntary participation; (3) Active undergraduate clinical medicine interns currently undertaking hospital rotations.
Exclusion Criteria: Responses exhibiting logical inconsistencies, excessively short completion times (<2 min), extreme response patterns, invalid answers to open-ended questions, or duplicate submissions.
Sample size determination followed psychometric standards recommending 10–30 participants per scale item (26). For the 7 core Likert-scale items, this yielded a minimum target of 210 participants. We recruited 324 interns through convenience sampling, of whom 21 were excluded based on predefined criteria. The final analytical sample of 303 participants substantially exceeded the psychometric target, thereby providing robust statistical power for factor analysis and minimizing the risks of model overfitting and factor instability.
2.4 Data analysis
Statistical analyses were performed using SPSS 25.0 and AMOS 26.0. All inferential tests employed α = 0.05 as the significance threshold. Descriptive statistics summarized participants’ overall characteristics, knowledge, and attitude scales. Qualitative data are described as frequencies and percentages; quantitative data are described as mean and Standard Deviation (SD). The scale items in Parts II and III underwent reliability and validity validation, employing Cronbach’s α coefficient to assess internal consistency, KMO measure and Bartlett’s test of sphericity for EFA, and AMOS 26.0 for Confirmatory Factor Analysis (CFA). Normality was assessed using the nonparametric Kolmogorov–Smirnov (K-S) test across gender, age, and city-tier subgroups. For normally distributed data, independent samples t-tests were employed; for non-normally distributed data, Mann–Whitney U tests (two-group comparisons) or Kruskal-Wallis tests (multi-group comparisons) (27)were utilized for mean comparisons. Spearman’s rank-order correlation analysis was performed to assess the relationship between knowledge and attitude dimensions. For Part III multiple-choice questions, multiple response analysis calculated option frequencies, with chi-square goodness-of-fit tests examining proportional differences across options and chi-square tests comparing distributions between subgroups. Part IV open-ended responses underwent a directed content analysis (28) using a predefined coding framework. Two doctoral researchers, independent of the study design, performed independent dual coding of all responses, a process totaling 25 h (mean: 5 min/response) across three consensus sessions over 10 days. This process achieved excellent inter-rater reliabilityκ = 0.82(κ > 0.80 represents almost perfect agreement beyond chance (29)). Coding discrepancies (14% of initial codes) were resolved systematically: consensus discussions resolved 85% of divergences, with unresolved cases arbitrated by the principal investigator using predefined decision rules. Final thematic frequencies were quantified to determine prevalence rates, supplemented with representative quotations to illustrate thematic essence while maintaining respondent anonymity.
2.5 Ethical and consent approval
This study was reviewed and approved by the Ethics Committee of Hangzhou Third People’s Hospital (Approval Number: 2025KA064). The committee granted a waiver of written informed consent as the research involved no more than minimal risk to participants, and the study utilized fully anonymized data collection with no personally identifiable information collected. In accordance with the ethical principles of the Declaration of Helsinki, comprehensive information regarding the research objectives, data handling procedures, and participant rights was prominently provided on the first page of the online questionnaire (Supplementary Material 1). Participants were explicitly informed that proceeding to and submitting the questionnaire would constitute their implied consent to participate in the study. This consent procedure was approved by the ethics committee.
3 Results
3.1 Sociodemographic characteristics of interns
CLKAQ questionnaires were distributed via the WeChat-based Questionnaire Star mini-program, yielding 324 submitted responses. After data screening excluding 21 invalid entries, 303 valid questionnaires were retained for analysis. The cohort comprised 75.2% interns aged 23–25 years and 24.8% aged 20–22 years. Gender distribution was 55.4% female and 44.6% male. Geographically, 45.2% represented tier-1 cities, 27.7% tier-2 cities, and 27.1% tier-3 cities.
Descriptive analysis of interns’ knowledge and attitude scales revealed distinct characteristics (Table 1). Based on 5-point Likert scoring, the 303 participants exhibited relatively low self-perceived knowledge scale scores (Q1-Q4; mean 2.22 ± 0.42), positioning in the lower-medium range. Conversely, attitude scale scores were notably higher (mean 4.05 ± 0.31), falling within the upper-medium range. Standard deviations for all items remained below 1 (range: 0.31–0.94), indicating low dispersion. Response distributions showed distinct clustering: knowledge items predominantly between Disagree and Neutral, while attitude items concentrated between Agree and Strongly Agree.
3.2 Scale outcomes by intern characteristics
Normality testing using K-S test across gender, age, and city-tier subgroups revealed non-normal distributions for all knowledge and attitude scale items (p < 0.001). Given the non-normality, the Mann–Whitney U test assessed differences between gender groups (male/female) and age cohorts (20–22/23–25 years), while the Kruskal-Wallis H test evaluated differences across city tiers (tier-1/tier-2/tier-3).
Comparative analysis of gender differences across scale items revealed statistically significant disparities on Q3, Q5, and Q7 (p < 0.05), with males demonstrating superior capabilities in independently interpreting laboratory reports (Q3: Z = −2.213), stronger endorsement of clinical laboratory knowledge’s critical role in clinical decision-making (Q5: Z = −5.267, p < 0.001), and heightened motivation to acquire additional knowledge (Q7: Z = −2.279) compared to female counterparts.
Analysis of age-group differences revealed statistically significant disparities on Q1, Q5, and Q7 (p < 0.05), with the 20–22 cohort demonstrating higher self-perceived adequacy of clinical laboratory knowledge for job requirements (Q1: Z = −3.153) compared to the 23–25 cohort, while the older group exhibited more positive attitudes regarding the critical role of laboratory knowledge in clinical decision-making (Q5: Z = −1.969) and stronger motivation to acquire additional knowledge (Q7: Z = −1.995).
Comparative Analysis Across City Tiers.
As shown in Table 4, interns from tier-3 cities exhibited significantly higher mean ranks than those in tier-1 and tier-2 cities across all knowledge (Q1-Q4) and attitude (Q5-Q7) scale items (p < 0.05), indicating stronger learning demands and more positive attitudes toward clinical laboratory medicine.
3.3 Analysis of multiple-choice responses on clinical laboratory attitudes
Significant variations were observed across all multiple-choice factors (p < 0.001; Table 5). For knowledge preferences (Q8), interpretation of routine tests (Q8a) and clinical-diagnostic integration (Q8b) demonstrated the highest demand, followed by specialized laboratory disciplines (Q8e), while technical principles (Q8c) and quality control (Q8d) were less prioritized. Motivations for learning (Q9) showed that 98.1% of interns prioritized clinical decision-making enhancement (Q9a), career advancement (Q9c), and certification exam preparation (Q9b), with only 1.9% Response Rate (n = 17) citing Intrinsic interest (Q9d). Knowledge acquisition channels (Q10) were dominated by Internet resources (Q10f), Medical literature (Q10a), Hospital-based training (Q10b), and Academic lectures (Q10c) (collectively 88.3% Response Rate), whereas mentor guidance (Q10e) and online courses (Q10d) accounted for merely 11.8% Response Rate. Primary learning barriers (Q11) included theory-practice disconnect (Q11a) and lack of expert guidance (Q11d) (80.9% Response Rate), with secondary barriers—Insufficient resources (Q11c), time constraints (Q11b), and content complexity (Q11e)—comprising 19.2% (Full multiple-choice response data are available in Supplementary Table S1).
Chi-square analyses further identified (Statistical details for non-significant comparisons are provided in Supplementary Table S2): Gender-based disparities in perceived learning barriers (Q11: χ2 = 13.051, p = 0.023), City-tier variations in knowledge acquisition channels (Q10: χ2 = 55.872, p < 0.001). These patterns are visualized in Figures 1, 2. Q11a was a cross-gender core option (cumulative percentage >80%), selected by 119 males (88.1%) and 152 females (90.5%); males demonstrated a greater inclination toward Q11d, with 64 selections (27.95%) compared to females’ 53 (21.12%); males showed the lowest selection rate for Q11e (5.2%), while females had a higher rate for this option (8.3%). Regarding cumulative coverage comparison: the top two options (Q11a + Q11d) accounted for 79.92% of male responses, with males selecting approximately 1.7 options on average; for females, the top two options accounted for 81.68%, with females selecting approximately 1.5 options on average (251 choices among 168 females), indicating a greater tendency among males to select multiple options.

Figure 1. Pareto analysis of Q11a option distribution by gender. N = total participants; n = number of respondents selecting the option.

Figure 2. Pareto analysis of knowledge acquisition channel preferences distribution by city tier: Q10 item preferences. N = total participants; n = number of respondents selecting the option.
Across all city tiers, Q10f was the most prevalent option, selected by 132 individuals (96.35%) in tier-1 cities, 81 (96.43%) in tier-2 cities, and 78 (95.12%) in tier-3 cities. Preference variations between tiers emerged: tier-3 cities exhibited more concentrated selections (the top two options accounting for 58.91%), whereas tier-1 cities showed a relatively balanced distribution (the top two options accounting for 48.76%); Q10a displayed an ascending acceptance gradient from tier-1 to tier-3 cities (from 23.42 to 28.68%), indicating a potentially greater focus on medical textbooks/literature in smaller cities; Q10d remained the option with the lowest proportion across all city tiers. According to the significance under Pareto Principle (30), the top three options covered 69.30% in tier-1 cities, 70.13% in tier-2 cities, and 80.23% in tier-3 cities — confirming that options (Q10f/Q10a/Q10b) satisfy the majority of interns.
3.4 Open-ended responses regarding attitudes toward testing among undergraduate clinical medicine interns
All 303 interns (100%) endorsed the need for enhanced clinical laboratory medicine education (Q12). Subsequent qualitative analysis of Q13 yielded detailed suggestions for improvement. Directed content analysis of these responses generated 65 distinct codes, which were iteratively synthesized into 5 overarching themes (Mandatory Clinical Laboratory Rotations, Test Report Interpretation Workshops, Hospital Academic Lectures on Emerging Technologies, Monthly Microlearning Online Modules and Case-Based Diagnostic Sessions) outlining proposed training strategies. The identified themes, presented in Table 6, encompass a range of practical and formalized training approaches, reflecting interns’ preferences for diverse and structured educational formats. The most prevalent recommendation was for mandatory clinical laboratory rotations (27.4%), highlighting a pressing need to gain a better understanding of laboratory operations and thereby acquire more knowledge about laboratory testing. These suggestions provide a clear framework for refining laboratory medicine training during clinical clerkships (Table 6).

Table 6. Themes from qualitative analysis of interns’ suggestions (Q13) for improving clinical laboratory medicine training.
4 Discussion
This study aimed to assess the self-perceived knowledge and attitudes toward clinical laboratory medicine among undergraduate medical interns in China, a population and topic that has not been comprehensively evaluated previously. Our findings reveal a significant discrepancy between interns’ modest self-assessed knowledge and their highly positive learning attitudes, alongside identifying specific gaps influenced by gender, age, and city tier. Furthermore, the study delineates preferred learning modalities and perceived barriers, providing a evidence-based foundation for targeted educational interventions. Since Dr. Gottfried’s (31) survey on establishing Laboratory Medicine courses in medical schools and publications by Smith (32) and Dr. Wilson (33) emphasized the significance of Laboratory Medicine knowledge, international attention to such curricula has gradually increased. Subsequently, Saffar et al. (10) designed an assessment of undergraduate medical students’ knowledge in Laboratory Medicine; however, this study had a relatively small sample size of 37 interns and employed a non-scale methodology. Professor Alosaimi (11) at Kingdom of Saudi Arabia and Bengayed et al. (34) in Tunisia, respectively, developed knowledge assessment questionnaires targeting specialized domains within Clinical Laboratory Medicine, aiming to underscore the critical importance of such knowledge. No study has comprehensively assessed self-perceived knowledge-attitude discrepancies among Chinese clinical interns. Building upon this foundation, we analyze the current challenges through our self-developed questionnaire (CLKAQ) and propose evidence-based improvements.
4.1 Knowledge-attitude discrepancy
This study identified a significant discrepancy between interns’ self-perceived knowledge of clinical laboratory medicine (mean score 2.22/5) and their proactive learning attitudes (mean score 4.05/5). This pattern reflects a hierarchical competence gap within Miller’s clinical assessment framework (35): while interns demonstrated strong motivation (action level), their operational capability in core competencies—exemplified by the low score in independently interpreting laboratory reports (Q3: 2.22; shows how level)—remained critically underdeveloped. Notably, Q3 recorded the lowest score (2.22), with only 20.8% of interns expressing confidence in interpreting basic reports. This highlights a critical deficit relative to the ACGME’s mandated competency of integrating laboratory results into clinical decision-making (1).
The predominant barriers to learning were a “curriculum-practice disconnect” (Q11a, 89.4%) and a “lack of expert guidance” (Q11d, 38.6%), underscoring a structural misalignment between laboratory medicine instruction and clinical practice requirements. Although a statistically significant positive correlation emerged between knowledge and attitudes (r = 0.171, p = 0.003), its weak magnitude suggests limited predictive utility of attitudes for knowledge levels. Nevertheless, the coexistence of a persistent knowledge gap and highly positive attitudes independently justifies targeted educational interventions.
This finding aligns with global studies reporting inadequate laboratory knowledge among medical trainees. For instance, Saffar et al. (10) found that 47.9% of Iranian medical students self-rated as ‘weak’ in test interpretation, consistent with our participants’ low confidence. Similarly, Alosaimi (11) reported that only 18.3% of Saudi interns passed basic microbiology assessments, reinforcing the widespread nature of this challenge. However, our study adds a crucial dimension by quantifying the acute awareness of this deficiency among Chinese interns, as evidenced by an overwhelming desire for enhanced knowledge acquisition (Q7, 97.7%)—an aspect not extensively detailed in previous international literature. The weak correlation further nuances this relationship, indicating that positive attitudes alone are insufficient to bridge knowledge gaps, necessitating structured educational reforms beyond mere awareness raising. Given the increasing complexity of laboratory testing, integrating practical laboratory medicine content into undergraduate medical curricula is more imperative than ever, a urgency underscored by Christian’s call for institutional reform (36).
4.2 Core issue of demand-resource mismatch
Responses to Q8 (“What types of clinical laboratory medicine knowledge do you most want to learn?”) reveal interns’ greatest needs lie in clinical-translational skills: Routine test interpretation (Q8a), Diagnostic integration (Q8b) and Specialized disciplines (Q8e). Conversely, Technical principles (Q8c) and Quality control and error analysis in diagnostic testing (Q8d) showed lower demand, indicating a need to reduce didactic instruction hours while increasing clinical scenario training. This aligns with the conclusions from Dr. Lawson’s (37) study: Laboratory sessions provide students opportunities to revisit concepts initially presented in traditional classroom settings and actively apply these concepts to case-based scenarios.
Responses to Q9 (“What are your primary motivations for learning clinical laboratory medicine?”) highlight practical drivers: enhancing Clinical decision-making (Q9a), Career development (Q9c), and Examination preparation (Q9b) – collectively emphasizing the applied utility of such knowledge. This strong external motivation underscores the perceived instrumental value of laboratory knowledge, a finding that corroborates the observations of Salinas et al. (38), who noted that since laboratory data informs 70% of clinical decisions, clinicians must be equipped to actively participate in these decision-making processes.
Analysis of Q10 (“How do you typically access clinical laboratory medicine knowledge?”) indicates interns primarily utilize Internet resources (Q10f, 96%), Medical literature (Q10a), Hospital-based training (Q10b), and Academic lectures (Q10c). Optimizing the quality and accessibility of these channels requires prioritized interventions: enhancing the authority and utility of Internet resources (Q10f) through clinical databases and guideline updates; integrating medical texts (Q10a) with digital platforms; and institutionalizing hospital training (Q10b) and academic lectures (Q10c). Notably, heavy reliance on unstructured learning via internet resources (Q10f) coexists with limited mentorship (Q10e, 28.7%). This self-directed paradigm exacerbates knowledge fragmentation, fundamentally undermining the core objective of “enhancing clinical judgment capabilities” (Q9a, 98.7%) and signaling critical resource misalignment. Consequently, elevating and formalizing the role of mentorship is imperative. This is strongly supported by evidence indicating that effective mentoring significantly enhances medical interns’ job satisfaction and aids in personal and professional career planning (39). Mentors should be equipped to deliver foundational medical knowledge more comprehensively, while medical education systems must innovatively leverage mentor expertise to create structured programs (40). Such structured mentorship not only enhances career selection confidence, increases scholarly output, and inspires careers in academic medicine (41), but also helps address challenges in mentor-mentee relationships, particularly for female interns, thereby demonstrably advancing equity across socioeconomic groups in medical training.
Analysis of Q11 (“What do you perceive as the main barriers to learning clinical laboratory medicine? “) identifies the foremost barrier as Theory-practice disconnect (Q11a), making curricular reform an immediate priority (as addressed in Section 4.1). Medical schools and residency programs should additionally incorporate interprofessional education, formal management training, and diversity-responsive teaching modalities (42). The secondary barrier—Inadequate expert mentorship (Q11d)—further underscores the critical significance of structured mentorship.
4.3 Educational implications of group disparities
Table 4 demonstrate that males exhibit greater confidence in report interpretation (Q3, p = 0.034), yet attitude scales indicate stronger learning motivation among females (Q7). This suggests males require enhanced training in knowledge rigor, while females benefit from increased practical confidence. Interns in tier-3 cities exhibit statistically significant higher mean rank values across knowledge and attitude scales compared to tier-1/2 cities, reflecting heightened engagement that may stem from primary healthcare settings’ heavier reliance on foundational testing. Notably, observed variation in responses to Q10 (“How do you typically access clinical laboratory medicine knowledge?”) across city tiers potentially stems from divergent learning environmental factors — a finding further supported by research emphasizing the practical implications of contextual learning influences, even in different trainee populations (43).
4.4 Synthesis of open-ended responses
The high utility of open-ended questions in uncovering respondent attitudes (44) is validated in this study, revealing undergraduate clinical interns’ critical demand for structured, specialized practical training in clinical laboratory medicine. The most endorsed recommendation was “Practical Rotation in the Testing Department” (27.4%), requiring ≥2-week immersive training to master key competencies including quality influencing factors and specialized fields like molecular biology. This reflects both the stringent operational standardization demanded in laboratory medicine (per CLSI EP23-A (45)) and the core need to enhance diagnostic accuracy through hands-on training – constituting a direct solution to the “knowledge-practice disconnect.” Complementary quarterly needs emerged for test report interpretation training (21.1%) and hospital academic lectures (20.1%). The former prioritizes laboratory physician-led interdisciplinary collaboration, addressing critical factors in translating laboratory data into clinical decisions. Diagnostic effectiveness fundamentally depends on two components: clinicians’ meticulous data analysis and recognition of laboratory medicine’s significance throughout healthcare processes. As David (9) notes, research on diagnostic errors remains critically inadequate—progress demands urgent methodological reform. The latter focuses on emerging technologies like molecular diagnostics. Together, they establish a longitudinal “theory-to-practice” training continuum. Notably, while online training (19.8%) is valued for flexibility, participants specified fragmented microlearning parameters (≤1 h monthly sessions), confirming trainee preference for approaches proven to enhance knowledge retention and reinforce long-term memory (46). Though case analysis received relatively lower endorsement (10.9%), its essential requirement for authentic patient data integration is designed to cultivate clinical correlation thinking.
4.5 Recommendations for educational improvement
Consistent with the study’s objective to inform curriculum optimization and educational strategy development, the following evidence-based recommendations are proposed based on the identified gaps in self-perceived knowledge, positive attitudes, and learning preferences.
Curriculum Restructuring with Dual-Track Laboratory-Clinical Integration: Implement longitudinal integration using clinical cases to sequence laboratory test selection → report interpretation → diagnostic refinement (e.g., quarterly case workshops per Q13 feedback). Embed mandatory laboratory rotations (27.4% support rate) within clinical clerkships, requiring ≥2-week placements emphasizing critical value management (Q4 mean score: 2.23) and hands-on experience with pre-analytical factors and basic instrumentation. This directly addresses the identified “knowledge-practice disconnect” (Q11a).
Resource Ecosystem Optimization: Develop authoritative, accessible knowledge repositories by curating and integrating high-quality Internet resources (Q10f, 96%) and medical texts (Q10a, 89.4%) with evidence-based guidelines (e.g., CLSI EP23-A). Establish structured mentor-trainee matching systems involving laboratory physicians to address the critical guidance deficiency (Q11d, 38.6%) through regular, laboratory-led report interpretation sessions (Q13, 21.1%). Effective mentoring is crucial not only in specialized residencies but also in undergraduate medical education to enhance confidence, clinical reasoning, and professional development.
Technology-Enabled Learning Innovation: The overwhelming reliance on Internet resources (Q10f, 96%) for self-directed learning, coupled with the identified barriers of “theory-practice disconnect” (Q11a) and “lack of expert guidance” (Q11d), underscores a critical strategic imperative: to transform fragmented information consumption into structured, effective education. We propose a multi-faceted digital strategy: First, develop AI-powered, adaptive learning platforms that curate authoritative information and provide personalized, on-demand guidance, effectively simulating the expert mentorship that is physically lacking. This directly addresses the accessibility and guidance gaps. Second, institutional investment in high-quality, gamified interactive modules is crucial. Evidence demonstrates that such modules significantly improve knowledge retention, engagement, and motivation in laboratory medicine education (16). These tools can bridge the theory-practice gap by immersing students in virtual clinical scenarios for test selection and interpretation.
5 Limitations and future directions
The study’s exclusive focus on hospitals in China may limit generalizability. Future research should validate the questionnaire’s external validity through multi-regional sampling. Uncontrolled confounding factors (e.g., internship duration, rotation departments) warrant stratified analysis or covariate adjustment in subsequent study designs. Beyond knowledge and attitudes, variables like career prospects, compensation structures, and professional development pathways may significantly influence learning outcomes; thus, their potential interaction effects should be investigated. A key limitation is the reliance on self-perceived knowledge rather than objective assessment of factual knowledge, which may introduce bias. Future studies should incorporate objective knowledge tests. Reliance solely on self-assessment of knowledge levels (without standardized assessments) necessitates future integration of objective knowledge evaluations to enhance the questionnaire’s real-world problem-reflecting capacity.
We anticipate that intensified focus on medical education reform and clinician training will drive continuous optimization of undergraduate clinical medical internships to address evolving educational demands.
6 Conclusion
This study provides the first comprehensive assessment of self-perceived knowledge and attitudes toward clinical laboratory medicine among undergraduate medical interns in China. The findings reveal a critical discrepancy between markedly positive learning attitudes (mean score 4.05/5) and deficient self-assessed knowledge (mean score 2.22/5), particularly in core competencies like independent report interpretation. This knowledge-attitude gap, exacerbated by a pronounced theory-practice disconnect and inadequate mentorship, underscores a systemic vulnerability in current clinical training paradigms.
The implications of these findings extend beyond identifying a problem; they offer a data-driven blueprint for educational reform. The strong, expressed demand for practical, immersive training—quantified through both quantitative and qualitative data—mandates a shift from theoretical didactics to integrated, experiential learning. Our evidence-based recommendations, including mandatory laboratory rotations, structured mentorship programs, and technology-enabled learning innovations, provide concrete pathways to achieve this shift. By addressing these specific gaps, medical educators can directly enhance interns’ diagnostic reasoning capabilities, ultimately contributing to improved patient safety and reduction in diagnostic errors.
The CLKAQ, rigorously developed and validated in this study, serves as a reliable tool for ongoing evaluation and benchmarking in this field. While the study’s focus on Eastern China may limit immediate generalizability, the methodological framework and identified themes offer valuable insights for other medical education systems facing similar challenges.
Future research should prioritize multi-institutional and longitudinal studies to validate these findings across diverse regions and to assess the long-term impact of the proposed interventions. Integrating objective knowledge assessments with self-perceived measures will further refine our understanding of these competencies.
In summary, this research confirms that empowering future physicians with robust laboratory medicine skills requires more than motivation—it demands structural educational change. By implementing the proposed strategies, medical education systems can bridge the identified chasm between enthusiasm and competency, thereby preparing a more clinically proficient generation of physicians.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.
Ethics statement
The studies involving humans were approved by the Ethics Committee of Hangzhou Third People’s Hospital (Approval Number: 2025KA064). The studies were conducted in accordance with the local legislation and institutional requirements. The ethics committee/institutional review board waived the requirement of written informed consent for participation from the participants or the participants’ legal guardians/next of kin because the questionnaire was distributed via QR code through WeChat, with respondents providing informed consent before participation. Participants could withdraw from the survey at any time if they disagreed. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.
Author contributions
YY: Data curation, Conceptualization, Writing – review & editing, Writing – original draft. JT: Conceptualization, Methodology, Writing – review & editing. BC: Writing – original draft, Methodology, Validation. SC: Resources, Formal analysis, Writing – review & editing, Software.
Funding
The author(s) declare that no financial support was received for the research and/or publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The authors declare that no Gen AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fmed.2025.1671631/full#supplementary-material
Abbreviations
CLKAQ, Clinical Laboratory Knowledge and Attitudes Questionnaire; K-S, Kolmogorov–Smirnov; SD, Standard Deviation; ICC, Intraclass Correlation Coefficients; I-CVI, Item-level Content Validity Index; CVR, Content Validity Ratio; EFA, Exploratory Factor Analysis; CFA, Confirmatory Factor Analysis; RMSEA, Root Mean Square Error of Approximation; GFI, Goodness-of-Fit Index; NFI, Normed Fit Index; CFI, Comparative Fit Index; TLI, Tucker-Lewis index; KMO, Kaiser-Meyer-Olkin; HPE, Health Professions’ Education; ACGME, Accreditation Council for Graduate Medical Education.
References
1. Accreditation Council for Graduate Medical Education (ACGME). Common Program Requirements (Residency). (2023) Available online at:https://www.acgme.org/globalassets/pfassets/programrequirements/cprresidency_2023.pdf
2. Abdollahi, A, and Jalali Nadoushan, MR. Little Attention Paid to Laboratory Medicine in Medical School Curriculum. Iran J Pathol. (2011) 6:201–1.
3. Yanfang, GENG, Miao, LI, and Xin, ZHAO. The Practice and Thinking of Teaching Status and Teaching Management Against Clinical Interns. China Continuing Med Education. (2021) 13:102–6. doi: 10.3969/j.issn.1674-9308.2021.25.027
4. Yangdong, FAN, Yanling, ZHU, Qing, ZHANG, Xinshu, ZHOU, Danping, XIE, and Pinhong, BAO. Suggestion on the Problems of Standardized Training of Residents in Guangzhou. Chinese Health Serv Manag. (2020) 37:58–61.
5. Libing, HONG. Analyses and suggestions on resident doctors standard training pattern in Hangzhou. Master's Degree: Central South University (2014).
6. Yun-wu, ZHAO, Heng, WANG, Xu-lin, CHEN, Quan, LI, Nian-nian, LI, and Shu-ling, CHEN. Analysis on the Problems of Standardized Training of Residents in China and Countermeasures. Health Econ Res. (2019) 36:12–4. doi: 10.14055/j.cnki.33-1056/f.2019.03.004
7. Working Committee for the Accreditation of Medical Education, Ministry of Education, P. R. China. Standards for basic medical education in China. Beijing: Peking University Press (2022).
8. Holmboe, ES, and Iobst, WF. Assessment guidebook. Accreditation Council for Graduate Medical Education. (2020), https://www.acgme.org/globalassets/pdfs/milestones/guidebooks/assessmentguidebook.pdf
9. Newman-Toker, DE, Nassery, N, Schaffer, AC, Yu-Moe, CW, Clemens, GD, Wang, Z, et al. Burden of serious harms from diagnostic error in the USA. BMJ Qual Saf. (2024) 33:109–20. doi: 10.1136/bmjqs-2021-014130
10. Saffar, H, Saatchi, M, Sadeghi, A, Asadi Amoli, F, Tavangar, SM, Shirani, F, et al. Knowledge of Laboratory Medicine in Medical Students: Is It Sufficient? Iran J Pathol. (2020) 15:61–5. doi: 10.30699/ijp.2020.94221.1916
11. Alosaimi, B, Saeed, AA, Mustafa, AA, AlJabr, WA, Batarfi, MA, and Hamza, MA. Recall of Prior Knowledge in Medical Microbiology Among Medical Interns: A Multicenter Cross-Sectional Assessment in Saudi Arabia. Adv Med Educ Pract. (2022) 13:733–9. doi: 10.2147/AMEP.S364330
12. Twisha, R. Banker, Marianne H Gillam, Peter O’Loughlin, Wayne Rankin, Richard Ryan, Connie Caruso, Elizabeth E Roughead, To test or to not test: A retrospective cross-sectional study on potentially inappropriate use of pathology testing in South Australian hospitals. Am J Clin Pathol. (2024) 161:342–8. doi: 10.1093/ajcp/aqad153
13. CNAS-CL02:2023 Accreditation Criteria for the Quality and Competence of Medical Laboratories. (2023). Available online at: https://www.cnas.org.cn/cms_files/filemanager/328304128/attach/202412/DE06AE9E831278695213E0EE4BF1AFBB.pdf.
14. Kearney, N, and Joy, A. Undergraduate dermatology education in Ireland: a survey of interns. Ir J Med Sci 1971. (2025) 194:107–15. doi: 10.1007/s11845-024-03842-y
15. Do, M, Sanford, K, Roseff, S, Hovaguimian, A, Besche, H, and Fischer, K. Gamified versus non-gamified online educational modules for teaching clinical laboratory medicine to first-year medical students at a large allopathic medical school in the united states. BMC Med Educ. (2023) 23:959. doi: 10.1186/s12909-023-04951-5
16. Artino, AR, La Rochelle, JS, Dezee, KJ, and Gehlbach, H. Developing questionnaires for educational research: AMEE Guide No. 87. Med Teach. (2014) 36:463–74. doi: 10.3109/0142159X.2014.889814
17. Tavakol, M, and Dennick, R. Making sense of Cronbach’s alpha. Int J Med Educ. (2011) 2:53–5. doi: 10.5116/ijme.4dfb.8dfd
18. Nguyen, NM, Tran, HTB, Do, TQN, Fait, P, and Reychler, G. The cross-cultured validity and reliability of the vietnamese nijmegen questionnaire. Cureus. (2025) 17:e78709. doi: 10.7759/cureus.78709
19. Koo, TK, and Li, MY. A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J Chiropr Med. (2016) 15:155–63. doi: 10.1016/j.jcm.2016.02.012
20. Wilson, FR, Pan, W, and Schumsky, DA. Recalculation of the critical values for Lawshe’s content validity ratio. Meas Eval Couns Dev. (2012) 45:197–210. doi: 10.1016/j.dza.2014.02.004
21. Polit, DF, Beck, CT, and Owen, SV. Is the CVI an acceptable indicator of content validity? Appraisal and recommendations. Res Nurs Health. (2007) 30:459–67. doi: 10.1002/nur.20199
22. Kaiser, HF, and Michael, WB. Little Jiffy factor scores and domain validities. Educ Psychol Meas. (1977) 37:363–5. doi: 10.1177/001316447703700210
23. Bartlett, MS. Tests of significance in factor analysis. Br J Stat Psychol. (1950) 3:77–85. doi: 10.1111/j.2044-8317.1950.tb00285.x
24. Hu, L, and Bentler, PM. Cutoff criteria for fit indexes in covariance structure analysis: conventional criteria versus new alternatives. Struct Equ Modeling. (1999) 6:1–55. doi: 10.1080/10705519909540118
25. Mac Callum, RC, Browne, MW, and Sugawara, HM. Power analysis and determination of sample size for covariance structure modeling. Psychol Methods. (1996) 1:130. doi: 10.1037/1082-989X.1.2.130
26. DeVellis, RF, and Thorpe, CT. Scale development: theory and applications. 5th ed. SAGE:, Thousand Oaks (2022).
27. Hazra, A, and Gogtay, N. Biostatistics series module 3: Comparing groups: numerical variables. Indian J Dermatol. (2016) 61:251–60. doi: 10.4103/0019-5154.182416
28. Hsieh, HF, and Shannon, SE. Three approaches to qualitative content analysis. Qual Health Res. (2005) 15:1277–88. doi: 10.1177/1049732305276687
29. Landis, JR, and Koch, GG. The measurement of observer agreement for categorical data. Biometrics. (1977) 33:159–74.
30. Harvey, HB, and Sotardi, ST. The pareto principle. J Am Coll Radiol. (2018) 15:931. doi: 10.1016/j.jacr.2018.02.026
31. Gottfried, EL, Kamoun, M, and Burke, MD. Laboratory medicine education in united states medical schools. Am J Clin Pathol. (1993) 100:594–8. doi: 10.1093/ajcp/100.6.594
32. Smith, BR, Aguero-Rosenfeld, M, Anastasi, J, Baron, B, Berg, A, Bock, JL, et al. Educating medical students in laboratory medicine: A proposed curriculum. Am J Clin Pathol. (2010) 133:533–42. doi: 10.1309/AJCPQCT94SFERLNI
33. Wilson, ML. Educating medical students in laboratory medicine. Am J Clin Pathol. (2010) 133:525–8. doi: 10.1309/AJCPQIA4FUGMVHT8
34. Bengayed, K, and Hafien, A. Audit about the knowledge of healthcare personnel regarding transfusion: A multicenter study. Tunis Med jul. (2024) 102:415–418. doi: 10.62438/tunismed.v102i7.4988
35. Miller, GE. The assessment of clinical skills/competence/performance. Acad Med. (1990) 65:S63–7. doi: 10.1097/00001888-199009000-00045
36. Christian, RJ, Baccon, J, Knollmann-Ritschel, B, Elliott, K, Laposata, M, and Conran, RM. The need for laboratory medicine in the undergraduate medical education curriculum: A white paper from the association of pathology chairs. Med Sci Educ. (2023) 34:193–200. doi: 10.1007/s40670-023-01895-9
37. Lawson, LB, Lind, CM, Gibson, JW, and Bentrup, HZ. Do Voluntary Lab-Based Active Learning Sessions Impact Medical Student Knowledge Retention? Med Sci Educ. (2020) 30:823–31. doi: 10.1007/s40670-020-00956-7
38. Salinas, M, López-Garrigós, M, and Uris, J. Towards laboratory knowledge, not data, in 70% of clinical decision-making. What “knowledge management” can add to clinical practice? Clin Chem Lab Med CCLM. (2011) 49:1389–90. doi: 10.1515/CCLM.2011.636
39. Han, ER, Chung, EK, Oh, SA, Woo, YJ, and Hitchcock, MA. Mentoring experience and its effects on medical interns. Singapore Med J. (2014) 55:593–7. doi: 10.11622/smedj.2014157
40. Platz, J, and Hyman, N. Mentorship. Clin Colon Rectal Surg. (2013) 26:218–23. doi: 10.1055/s-0033-1356720
41. Murphy, AM, Shenot, PJ, and Lallas, CD. Faculty development: How do we encourage faculty to become better teachers and mentors? Curr Urol Rep. (2020) 21:40. doi: 10.1007/s11934-020-00994-z
42. Bai, H. Modernizing medical education through leadership development. Yale J Biol Med. (2020) 93:433–9.
43. Alhenaidi, A, Alqallaf, B, Alsalahi, S, Al-Haqan, A, Elsherif, M, and Kelendar, H. Association between learning environment and well-being of postgraduate residents in kuwait. BMC Med Educ. (2025) 25:457. doi: 10.1186/s12909-025-07030-z
44. Hansen, K, and Świderska, A. Integrating open- and closed-ended questions on attitudes towards outgroups with different methods of text analysis. Behav Res Methods. (2023) 56:4802–22. doi: 10.3758/s13428-023-02218-x
45. Clinical and Laboratory Standards Institute. Laboratory quality control based on risk management (2nd ed.): CLSI guideline EP23[S]. Wayne: CLSI. (2023).
Keywords: medical school curriculum, laboratory medicine knowledge, clinical medicine students, undergraduate medical education, medical interns
Citation: Yang Y, Tian J, Chen B and Chen S (2025) Knowledge and attitudes toward clinical laboratory medicine among undergraduate medical interns in China: a cross-sectional survey. Front. Med. 12:1671631. doi: 10.3389/fmed.2025.1671631
Edited by:
Goran Hauser, University of Rijeka, CroatiaReviewed by:
Abdel-Hady El-Gilany, Mansoura University, EgyptWeeratian Tawanwongsri, Walailak University, Thailand
Fariba Khanipoor, Shiraz University of Medical Sciences, Iran
Copyright © 2025 Yang, Tian, Chen and Chen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Yonggang Yang, NDM3NTQ5NjJAcXEuY29t
†ORCID: Yonggang Yang, orcid.org/0009-0009-7799-8829
Jiyun Tian, orcid.org/0000-0002-0472-3519