Your new experience awaits. Try the new design now and help us make it even better

REVIEW article

Front. Pediatr., 10 September 2025

Sec. Pediatric Critical Care

Volume 13 - 2025 | https://doi.org/10.3389/fped.2025.1636580

Predicting nosocomial infections in critically Ill children: a comprehensive systematic review of risk assessment models


Ling-Ying WangLing-Ying Wang1Mei FengMei Feng2Yu-Lan LuoYu-Lan Luo1Chun-Xia WangChun-Xia Wang1Heng WangHeng Wang3Li LiLi Li1Yuan ZhangYuan Zhang1Xiu-Ling HuangXiu-Ling Huang1Min-Jie HuangMin-Jie Huang1Yong-Ming Tian

Yong-Ming Tian1*
  • 1Department of Critical Care Medicine, West China Hospital, Sichuan University/West China School of Nursing, Sichuan University, Chengdu, China
  • 2Nursing Department, West China Hospital, Sichuan University/West China School of Nursing, Sichuan University, Chengdu, China
  • 3Department of Anesthesiology, West China Hospital, Sichuan University, Chengdu, China

Background: Nosocomial infections (NIs) pose a substantial global health challenge, affecting an estimated 7%–10% of hospitalized patients worldwide. Neonatal intensive care units (NICUs) are particularly vulnerable, with NIs representing a leading cause of infant morbidity and mortality. Similarly, pediatric intensive care units (PICUs) report that 28% of admitted children acquire NIs during hospitalization. Although prediction models offer a promising approach to identifying high-risk individuals, a systematic evaluation of existing models for ICU-ill children remains lacking.

Aim: This review systematically synthesizes and critically evaluates published prediction models for assessing NI risk in ill children in the ICU.

Methods: We conducted a comprehensive search of PubMed, Embase, Web of Science, CNKI, VIP, and Wanfang from inception through December 31, 2024. Study quality, risk of bias, and applicability were assessed using the PROBAST tool. Model performance metrics were extracted and summarized.

Results: Three studies involving 1,632 participants were included. Frequency analysis identified antibiotic use, birth weight, and indwelling catheters as the most consistently incorporated predictors. All models employed traditional logistic regression, with two undergoing external validation. However, critical limitations were observed across studies: inadequate sample sizes, omission of key methodological details, insufficient model specification, and a universally high risk of bias per PROBAST assessment.

Conclusion: Current NI prediction models for ill children in the ICU exhibit significant methodological shortcomings, limiting their clinical applicability. No existing model demonstrates sufficient rigor for routine implementation. High-performance predictive models can assist clinical nursing staff in the early identification of high-risk populations for NIs, enabling proactive interventions to reduce infection rates. Future research should prioritize (1) methodological robustness in model development, (2) external validation in diverse settings, and (3) exploration of advanced modeling techniques to optimize predictor selection. We strongly advocate adherence to TRIPOD guidelines to enhance predictive models' transparency, reproducibility, and clinical utility in this vulnerable population.

Systematic Review Registration: PROSPERO CRD420251019763.

1 Introduction

Over seven decades of development, critical care medicine has undergone transformative, industrial-scale advancements in managing life-threatening conditions. However, mirroring the unintended consequences of industrialization, these breakthroughs- particularly for vulnerable populations (e.g., elderly, immunocompromised, and chronically ill patients) - have inadvertently increased the prevalence and complexity of nosocomial infections (NIs) in modern intensive care units (ICUs) (1). The US Center for Disease Control and Prevention identifies that nearly 1.7 million hospitalized patients annually acquire NIs while being treated for other health issues and that more than 98,000 patients (one in 17) die due to these (2). The burden of NIs demonstrates marked geographical inequalities. Lower- and middle-income countries (LMICs) report an average NI prevalence of 17%, with surgical site infections constituting the predominant type, followed by urinary tract infections, bloodstream infections, then respiratory tract infections (3). In striking contrast, high-income countries (HICs) document NI rates approaching 30% among ICU patients - a phenomenon attributable primarily to three key factors: (1) higher prevalence of patient comorbidities, (2) greater use of immunosuppressive therapies, and (3) more frequent employment of invasive medical devices (3, 4).

NIs in neonatal intensive care units (NICUs) represent a particularly alarming global health concern, significantly contributing to infant morbidity and mortality worldwide (5). Premature neonates are confronted with a host of compounded vulnerabilities: prolonged hospitalizations, immature immune systems, fragile skin integrity, and frequent exposure to life-saving yet infection-prone medical interventions (57). While the United States witnessed significant reductions in central line-associated bloodstream infections (CLABSIs) between 2007 and 2012, recent progress in this area has unfortunately stagnated (8, 9). In contrast, in low- and middle-income countries (LMICs), neonatal nosocomial infections (NIs) impose a disproportionately higher burden, mainly due to resource constraints and care disparities (10). Globally, in pediatric intensive care units (PICUs), 28% of children develop nosocomial infections during their stay (11). In such critical care settings, the timely identification of high-risk child patients and the implementation of preventive interventions are of paramount importance. Complications such as prolonged length of stay in ICU, prolonged length of stay in hospital, excess hospitalization costs, and predicted mortality are highly associated with NIs (12, 13), and place a substantial economic burden on healthcare systems (14). Fortunately, modifiable care processes present actionable targets for reducing infection prevalence, offering hope for meaningful improvements in patient outcomes and healthcare efficiency.

In clinical settings, the current identification of NIs primarily relies on specimen culture-based diagnostics, which typically require several days to produce results. This delay often necessitates initiating empirical broad-spectrum antibiotic therapy, exposing non-infected patients to unnecessary treatments and exacerbating the selective pressure for antibiotic-resistant pathogens (15). Moreover, traditional NI risk stratification remains inherently subjective, as it depends on clinician experience and qualitative interpretations of patient data. This subjectivity leads to inconsistent decision-making and limited reproducibility across care providers (16). Such variability compromises the reliability of correlating clinical risk factors with prognostic outcomes, further underscoring the critical role of practitioner expertise in contextualizing patient-specific scenarios. Clinical prediction models (CPMs) offer a promising alternative by providing objective, algorithm-driven frameworks that integrate multidimensional risk factors into statistical models to estimate disease probability or event risk in defined populations (1719).

By leveraging historical patient data, data-driven decision support systems can effectively mitigate cognitive biases in clinical judgment, delivering standardized prognostic insights that enhance care precision (20). However, while numerous CPMs have been developed for NI risk prediction in adult ICUs, pediatric applications in neonatal and pediatric critical care populations remain significantly underdeveloped. Existing pediatric models often lack systematic evaluation, comparative benchmarking, or consensus-based validation for clinical implementation. Additionally, the methodological rigor across studies varies widely, necessitating rigorous appraisal to establish evidence-based recommendations for practice.

To address these gaps, this study conducts a comprehensive performance evaluation of all published, validated NI risk prediction models specific to ill children in ICUs. Through structured comparison and quality assessment, we aim to identify optimal predictive tools for clinical translation and formulate prioritized research directions to advance NI risk stratification in critically ill children.

2 Methods

2.1 Design

This study adhered to the Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies (CHARMS) checklist to systematically evaluate predictive modeling studies (21). The review was planned and reported by the PRISMA 2020 guidelines (22). The protocol was registered on the International Prospective Register of Systematic Reviews (PROSPERO, ID: CRD420251019763).

2.2 Search strategy

We conducted a comprehensive systematic search across six major databases: PubMed, Web of Science, Embase, China National Knowledge Infrastructure (CNKI), Chinese Technical Periodicals (VIP), and Wanfang databases, covering their inception to December 2024. The search strategy incorporated relevant medical subject headings (MeSH) and free-text terms related to ICU, infection control, and predictive modeling. A detailed description of the complete search strategy is provided in the Supplementary Table S1. Additionally, the reference lists of all included studies and relevant reviews were reviewed to identify any additional references.

2.3 Study eligibility criteria

We established the inclusion and exclusion criteria before the study.

2.3.1 Inclusion criteria

1. Study population: Children aged under 18 years.

2. Study content: Studies on risk prediction models for NIs in ICU child patients (including neonatal and pediatric populations).

3. Study type: Case-control studies and cohort studies.

2.3.2 Exclusion Criteria

1. Studies focusing on adult populations (age ≥ 18 years).

2. Non-longitudinal study designs.

3. Models with fewer than two predictors.

4. Duplicate publications.

5. Unofficial publications, such as conference abstracts and academic papers.

6. Studies not available in Chinese or English.

2.4 Study selection

We imported the search results into EndNote X9 software for data management. After removing duplicates, the titles and abstracts of the retrieved studies were screened. Full texts of potentially eligible studies were then obtained and evaluated. Two authors independently assessed all studies, with any disagreements resolved through discussion with additional review authors. Multiple publications from the same model were compiled, with the most comprehensive report designated as the primary reference.

2.5 Data extraction

Data extraction was independently conducted by two authors using Microsoft Excel, with discrepancies resolved through consultation with other review authors. Standardized data extraction forms were designed based on the CHARMS checklist (21). The critical information extracted followed the PICOTS principles, including the number of subjects included, data source, predictors (e.g., age, albumin infusion), model status (e.g., performance, modeling status, and model presentation), and outcome metrics. We collected information such as author names, year of publication, study type, and statistical details (e.g., treatment of missing data, selection of predictors, and treatment of continuous variables).

To analyze the predictive ability of each model, the following metrics were used to evaluate the clinical applicability:

1. Discrimination: The model's ability to distinguish between individuals with and without the outcome of interest, often measured by the Consistency Statistics (C-index) and the Area Under the Curve (AUC). The closer the AUC is to 1, the better the diagnostic effectiveness of the model (23);

2. Calibration: The accuracy of probability predictions, measured by the Hosmer-Lemeshow test and calibration curves (24);

3. Clinical Validity Evaluation Metrics: Decision Curve Analysis (DCA) was used to assess the clinical utility of predictive models, aligning with practical clinical decision-making processes (25). In addition to these metrics, the confusion matrix, accuracy, sensitivity, specificity, F1-score, and Brier score were also evaluated (26).

2.6 Quality assessment

We applied the Risk of Bias Assessment Tool (PROBAST) (27) to assess the risk of bias (ROB) and the applicability of the prediction models. PROBAST comprises four domains: participants, predictors, outcomes, and analysis. Each question can be answered as “yes,” “probably yes,” “probably no,” “no,” or “no information.” A domain was considered high risk if any question was answered “no” or “probably no.” Conversely, a domain was defined as low risk if all questions were answered “yes” or “probably yes.” The overall ROB was deemed low when each domain consistently exhibited a low ROB. If one or several domains exhibited an uncertain ROB while the remaining domains were low risk, the ROB was categorized as unclear. The applicability evaluation followed a similar approach, but only the first three domains were used to assess the applicability of the predictive model. Two authors independently evaluated the risk of bias in the included models, with any discrepancies resolved through consultation with other review authors.

2.7 Data synthesis and analysis

Given the significant heterogeneity in study design, populations, and outcomes across the studies, we decided against pooling data for meta-analysis. This decision ensured that the unique findings of each study, reflecting diverse study designs and population characteristics, were preserved.

3 Results

3.1 Study selection and characteristics

The PRISMA flowchart illustrates our process for searching and selecting literature (Figure 1). A total of 6,857 articles were identified in the initial search. After removing duplicates, 3,846 articles remained for title and abstract screening. Full-text reviews were conducted on 106 articles, of which only three were ultimately included in this review (2830).

Figure 1
Flowchart depicting the selection process for a systematic review. Initially, 6,857 records were identified from databases such as PubMed and Embase, with additional sources adding 2 more. After removing 3,013 duplicates, 3,846 records were screened, excluding 3,740 after title and abstract review. From 106 full-text articles assessed, 103 were excluded for reasons such as age criteria and relevance. Ultimately, 3 articles were included in the review.

Figure 1. Flowchart of the selection of studies for inclusion in the systematic review.

3.2 Characteristics of the included studies

The characteristics of the included studies are summarized in Table 1. The three studies were published between 2016 and 2023 and conducted across Asia, specifically in China and India.

Table 1
www.frontiersin.org

Table 1. Characteristics of the included studies and model development.

3.2.1 Study design and population

The modeling research designs were predominantly retrospective (28, 29), with one study employing a prospective design (30). The sample sizes of the two retrospective studies were 549 and 672, respectively, while the prospective study included 411 participants. One study focused on patients in the PICU aged 1 month to 12 years (30), while the other two involved NICU patients (28, 29).

3.2.2 Outcomes

The outcome measures in the three included studies were not identical. Zhou et al. (28) used multidrug-resistant organisms (MDROs) as the outcome measure for modeling. Miao et al. (29) focused on umbilical vein catheterization bloodstream infections, while Saptharishi L. G. et al. (30) used healthcare-associated infections (HAIs) as the outcome measure for modeling.

3.2.3 Predictors selection

All three studies used univariate analysis to select the most relevant factors for their models. A total of 17 final variables were identified across the studies, as shown in Table 1. Among the 17 predictors, three variables were most consistently incorporated across studies: antibiotic use (included in 2 studies): “use of antibiotics >7 days” (28) and “prior antibiotic use (≥4)” (30) were identified in both NICU and PICU models. Birth weight (included in 2 studies): “low birth weight” and “birth weight <1,500 g” (28, 29) were exclusive to NICU models. Indwelling catheters (included in 2 studies): “total indwelling time >7 days” (29) and “presence of indwelling catheters” (30) were featured in both NICU and PICU models.

Other predictors were included in only one study, including maternal age ≥35 years, MDRO colonization, serum albumin <35 g/L, puncture attempts >2 times, history of mechanical ventilation, history of peripherally inserted central catheter (PICC) treatment, age (<5 years), pediatric risk of mortality III (PRISM III) score (within 24 h), need for intubation, albumin infusion, and immunomodulator use, reflecting variability in study populations (NICU vs. PICU) and outcomes (MDROs vs. catheter-related infections vs. HAIs).

3.2.4 Missing data handling

In the study by Zhou et al., a total of 459 cases of nosocomial infections were detected in the modeling cohort. After excluding 24 cases (5.2%) with incomplete data, 435 neonates were finally included to construct the prediction model; the cases in the validation cohort had complete data, with no exclusions (28). In the study by Miao et al., the completeness of clinical data was set as an inclusion criterion. In the modeling group, 17 cases (3.7%) were excluded due to missing clinical and laboratory data, and 6 cases (2.6%) were excluded in the validation group (29). In the study by Saptharishi et al., among the 412 initially screened eligible subjects who signed the informed consent form, 1 case (0.2%) with incomplete data was excluded, and finally 411 cases were included in the model construction (30).

3.2.5 Modeling methods and validation

All three studies utilized multivariate logistic regression for modeling. Two studies conducted internal and external validation after model establishment (28, 29), while the other research combined model construction with internal validation using the Bootstrap method (30). Discrimination and calibration were reported for all three models, and only one model used DCA to assess clinical validity (28). The discrimination AUC values of the internally validated models ranged from 0.773 to 0.866.

3.2.6 Model presentation and reporting standards

Two of the models were presented as nomograms (28, 29), while the presentation format of the third model was scoring table (30). The Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) guidelines were published in 2015 to standardize the reporting of prediction model research, ensuring transparency and reproducibility (19, 31). Although all three included studies were published after 2015, none explicitly stated that they followed the TRIPOD statement. The completion status of the TRIPOD checklist among the three studies is shown in Supplementary Table S2.

3.2.7 Risk of bias and applicability

The PROBAST bias risk assessment results revealed that all three included models had a high risk of bias (Figure 2). Among the four assessed dimensions, the statistical analysis dimension had the worst assessment results, with all three models exhibiting a high risk of bias. Detailed scores for each dimension are shown in Supplementary Table S1.

Figure 2
Bar chart titled \

Figure 2. Risk of bias assessment for all included models (N = 3).

One study was deemed to have a high risk for applicability due to the lack of external validation (Figure 3).

Figure 3
Bar chart titled \

Figure 3. Applicability assessment for all included models (N = 3).

4 Discussion

This review comprehensively assessed the current state of predictive models designed to forecast NIs in ICU child patients. We identified three models and summarized their performance and the predictors used. Additionally, we conducted an in-depth analysis and identified several methodological shortcomings in developing and validating these predictive models.

All studies employed Logistic regression, a commonly used method for constructing models with binary classification variables, which likely reflects practical and contextual factors inherent to pediatric ICU research. On the one hand, Logistic regression generates odds ratios that are clinically intuitive, allowing clinicians to directly link individual predictors (e.g., “antibiotic use >7 days”) to infection risk (28, 29). This interpretability is particularly valued in neonatal and pediatric settings, where clinical decision-making often requires clear mechanistic links to justify interventions. On the other hand, the included studies had relatively small samples (411–672 participants), with event rates as low as 5.1% (29). Logistic regression is robust to small datasets, remains a default due to its low computational demand, and familiarity among researchers. And avoids overfitting compared to complex models (e.g., machine learning), which require larger samples to generalize (28). However, the issue of multicollinearity was not addressed in any of the models, which may compromise model performance. Machine learning (ML) techniques, such as random forest (RF), artificial neural networks (ANN), LASSO, and eXtreme gradient boosting (XGBoost), have demonstrated robust capabilities in handling high-dimensional data and strong generalization abilities (32). Therefore, in developing models to predict the impact of NIs in ICU child patients, it is advisable to explore alternative algorithms like RF, LASSO and XGBoost. Pediatric ICU data, especially for critically ill children, often suffer from small sample sizes and limited event rates (e.g., 5.1% for umbilical vein catheter-related infections (29), which constrain traditional statistical methods like logistic regression. ML techniques offer unique advantages in this context: Handling small datasets with regularization: (1) Algorithms such as LASSO regression and elastic net incorporate built-in regularization, which penalizes overfitting by shrinking coefficients of less relevant predictors. This is critical for pediatric data, where limited events (e.g., 34 infections in 672 NICU patients (29) increase the risk of overfitting with conventional methods. (2) Leveraging feature interactions without large samples: ML models (e.g., random forests, gradient boosting) automatically capture non-linear relationships and interactions between predictors (e.g., the combined effect of low birth weight and prolonged antibiotic use (28, 29) without requiring the large sample sizes needed to validate such interactions in regression models. This is valuable given the complex, multifactorial nature of pediatric NI risk. If the performance of models developed using a single algorithm is unsatisfactory, ensemble learning methods could be employed to create ensemble models. These models automatically model non-linear relationships and interactions without manual transformation, capturing complex patterns in pediatric data, integrate the strengths of multiple algorithms, significantly enhancing prediction accuracy (33).

Most studies directly deleted missing values, leading to incomplete reporting. This approach can introduce bias, as missing data can distort model performance if correlated with other variables (34). Missing value imputation methods include deletion, simple imputation, multiple imputation, and algorithmic imputation. Multiple imputation (35) and Miss Forest (36) are currently more recommended. About half of the studies also transformed continuous variables into binary or multi-class classifications. While this simplifies clinical decision-making, it can result in loss of data information and reduced prediction performance (37). The choice of method should be guided by the study's purpose, the data characteristics, and the intended application.

Three studies selected predictors using univariate analysis. However, with the advancement of modeling methods, predictor selection is crucial for model quality and effectiveness. Predictors should be chosen by combining clinical literature with data-driven approaches. Relying solely on statistical significance for predictor selection can lead to overfitting and overly optimistic performance measures (38). Variable selection methods can be categorized into filter, wrapper, and embedding (39). Regularized regression (penalized models or shrinkage methods) can help reduce overfitting for data with many features or multicollinearity. Machine learning algorithms, such as random forest, can also be used for predictor selection and dimensionality reduction (40), thereby improving research quality. The predictors in the models identified in this review varied substantially based on the outcomes and study population, demonstrating the multifactorial nature of risk factors associated with NIs in ICU child patients. For frequently occurring predictors (included in 2 studies): Antibiotic use reflecting the well-established link between prolonged antibiotic exposure and increased risk of NIs, particularly MDRO infections (28, 30). Birth weight highlights the critical role of immature immune function and fragile skin barrier in neonatal susceptibility to infections (5, 6). Indwelling catheters were featured in both NICU and PICU models, consistent with the known association between invasive devices and infection risk due to disruption of physical barriers and pathogen colonization (4, 41). Furthermore, the sample size is intrinsically linked to the variables under consideration. Beyond the conventional 10-EPV (events per variable) estimation approach, advanced sample size calculation tools have been specifically developed to more accurately estimate the required sample size for clinical prediction models (42).

Two studies conducted internal and external validation, while another combined model construction with internal validation. Calibration, which assesses whether the predicted outcome matches the observed outcome, was investigated for all models. Poorly calibrated models can mislead decision-making processes (43). Predictive models often perform better on development data than external validation data, but external validation is more convincing than internal validation (44). Therefore, models should be validated using different datasets whenever possible to ensure generalizability. Validation studies should verify that the model's performance on new data is comparable to its performance on the development data (45), and model usefulness should be assessed through clinical judgment.

The included studies exhibited inadequate model presentation and incomplete regression equations. Poor presentation wastes research resources and hinders future activities such as validation, updating, and clinical application. Various forms can be used for model presentation, including scoring systems, nomograms, web calculators, and mobile apps. In this study, two models were presented using nomograms, while one model adopted a scoring table. None of the included studies reported converting their models into bedside scoring systems, online calculators, or EHR-embedded alert formats critical for real-time decision-making in busy ICUs. None of the studies reported formal usability testing with clinical end-users (e.g., ICU nurses, physicians). The limited attention to model presentation and usability in the included studies reflects a common focus on methodological rigor (e.g., discrimination, calibration) over practical implementation in early-stage model development. The Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) guidelines, published in 2015 in multiple well-known journals (19, 31), aim to standardize the reporting of prediction model research to ensure transparency and reproducibility. Although all three included studies were published after 2015, none explicitly stated adherence to the TRIPOD statement. Adhering to TRIPOD guidelines can significantly enhance transparency and reproducibility. We strongly recommend that researchers submit the TRIPOD checklist when submitting manuscripts to facilitate evaluation by journal editors and reviewers. Full compliance with these standards will ensure scientific rigor in model development and improve the model's applicability in clinical settings.

Most studies developed new models with excellent predictive power. However, these models consistently exhibited a high risk of bias (ROB). Among the two studies (28, 29) based on a retrospective design, there was a high risk of bias in the assessment of study subjects. None of the three studies reported whether the predictors were assessed without knowledge of the outcome data or whether all predictors were available at the time the model was intended to be used, or whether the outcome was determined without knowledge of the predictor information. Thus, the risk of predictor and outcome bias was rated as “unclear.” All models had issues in statistical analysis, such as insufficient sample size, inadequate consideration of overfitting, handling of missing data, and unclear treatment of continuous variables. This can result in good performance on the training set but poor performance on the test set or in real-world applications (44). Overfitting means that while the model learns the training data well, it cannot generalize to new data effectively, significantly reducing its usefulness and reliability (46). Therefore, identifying child patients who could benefit from interventions to prevent NIs remains a critical public health strategy. Predictive models used with clinical decision support have been shown to improve patient outcomes (47) and should be considered when deploying risk models in ICU child patients.

Last but not least, it is essential to evaluate models' impact in practice. In our review, two studies were conducted in NICUs and one in a PICU, with populations from China and India only. The baseline infection rates, pathogen spectra, and resource availability in the study settings (China and India) exhibit distinct characteristics that may constrain the external validity of the models. The included studies were conducted in high-volume tertiary centers, where the burden of NIs may differ from other settings. For example, the NICU studies from China reported 16.2% (89/549) and 5.1% (34/672) infection rates for MDRO infections and umbilical vein catheter-related bloodstream infections, respectively (28, 29), while the Indian PICU study reported a HAI rate of 23.1% (95/411) (30). The models focus on region-specific pathogens: Zhou et al. (28) targeted MDROs, which are prevalent in Asian tertiary centers with high antibiotic usage; Miao et al. (29) focused on umbilical vein catheter-related pathogens, relevant to neonates in resource-intensive NICUs; and Saptharishi et al. (30) used HAI definitions aligned with CDC criteria but in an Indian PICU, where pathogen profiles may include more community-acquired or drug-resistant strains due to limited antimicrobial stewardship (11). The included models incorporate predictors tied to local resource access, such as “antibiotic use >7 days” (28), “mechanical ventilation history” (29), and “indwelling catheter presence” (30). Conversely, resource-rich HICs may prioritize different predictors (e.g., immunocompromised status, advanced organ support) not captured in the current models, further limiting cross-context applicability. Based on the available evidence, none of the identified models are currently validated for global use. Their applicability may be restricted to settings with similarities to the original study populations.

We additionally searched for studies that implemented NI prediction models in child patients in ICUs, but we could not identify any references. However, evaluating whether introducing a prediction model changes to care, e.g., increases interventions and improves outcomes, e.g., reduces the incidence of NIs, will be necessary for future investigations. Based on these methodological shortcomings, we make the following recommendations. First, models should be externally validated several times in different populations, and sample sizes must be adequately considered. Second, when data are missing, interpolation should be performed via multiple interpolations or machine learning. Third, predictive variables with incremental solid values should be mined based on clinical feasibility and applicability, and preventing overfitting should be emphasized in the predictive model. At the same time, we also need to recognize the importance of diversified evaluation, as far as possible, sensitivity, specificity, the calibration index, the net benefit, and DCA for comprehensive evaluation. Finally, model development is strongly recommended to adhere to the TRIPOD process. Only complete, full, and transparent reporting of all aspects of a prediction model can its risk of bias and potential usefulness be adequately assessed.

The distinction between NICU and PICU populations introduces critical variability that limits cross-setting generalizability: Developmental stage: NICU populations consist of neonates, often premature, with immature immune systems, fragile skin barriers, and prolonged exposure to invasive interventions (e.g., umbilical catheters) (5, 6). In contrast, PICU patients are older children (1 month to 12 years (30) with diverse comorbidities and varying degrees of immune competence. Risk factors: NICU models emphasize neonatal-specific predictors such as birth weight <1,500 g (29) and maternal age (28), which are irrelevant to older PICU patients. PICU models, meanwhile, focus on factors like intubation need and PRISM III scores (30), reflecting the acuity of critical illness in older children. Infection dynamics: Neonates in NICUs are disproportionately susceptible to device-related infections (e.g., umbilical vein catheter sepsis (29), whereas PICU patients face broader HAI risks linked to mechanical ventilation and immunomodulatory therapies (30). These differences mean that models developed for one setting (e.g., NICU) may not accurately identify high-risk patients in the other (e.g., PICU), as the underlying pathophysiology and risk profiles diverge.

Outcome variability further constrains generalizability: MDROs as a subset of HAIs: MDRO infections (28) represent a specific, drug-resistant subgroup of HAIs, driven by factors like prolonged antibiotic exposure (28, 30). This focus limits applicability to settings with high antibiotic stewardship or low MDRO prevalence. Device-specific vs. broad HAIs: The umbilical vein catheter-associated infections modeled by Miao et al. (29) are niche outcomes relevant only to neonates with invasive devices, whereas Saptharishi et al.'s (30) broader HAI definition (encompassing all healthcare-acquired infections) has a wider scope but may mask device-specific risks. Diagnostic criteria: Variations in diagnostic standards (e.g., CDC criteria for HAIs (30) vs. Chinese Ministry of Health criteria for MDROs (28) complicate cross-study comparisons and reduce the transferability of findings to settings with different surveillance protocols. Thus, models targeting MDROs or device-specific infections cannot be generalized to predict all HAIs, and vice versa, as their underlying etiologies and risk factors differ.

This review aimed to identify and evaluate predictive models for NIs in ICU child patients to inform clinical decision-making. However, we cannot recommend any specific model for several reasons. First, nearly all reviewed models exhibited a high risk of bias, and the included models require further external validation. Additionally, significant heterogeneity among the models, non-standardized statistical analysis methods, and incomplete data in model reports all contribute to the challenge of selecting the optimal model. While no existing model meets the criteria for routine clinical application, we propose the following interim strategies to address the urgent need for NI risk stratification in critically ill children: (1) Adapting validated adult ICU models with pediatric-specific adjustments: (A) Adult NI prediction models (e.g., those incorporating variables like invasive device duration, antibiotic exposure, and comorbidity burden) (4, 41) could serve as a foundation, with key modifications to account for pediatric physiology. (B) Adjusting for developmental factors [e.g., replacing adult age with postmenstrual age in neonates (5, 6)]. Incorporating pediatric-specific variables identified in our review, such as birth weight (28, 29), umbilical catheter use (29), and PRISM III scores (30). (C) Calibrating risk thresholds to reflect higher baseline vulnerability in children [e.g., lower thresholds for initiating interventions in NICUs due to immature immune function (7)]. (2) Implementing structured clinical risk assessment tools: Given the paucity of validated models, we recommend using consensus-based checklists that integrate the most consistent predictors from included studies: Antibiotic use exceeding 7 days (28, 30), presence of indwelling catheters (29, 30), and (in neonates) low birth weight (28, 29).

To address the “methodological robustness” gap highlighted in our conclusions, we specify the following priorities for future research, based on PROBAST and TRIPOD guidelines (19, 27). (1) Rigorous model development: (A) Sample size and event rates: Adhere to modern standards [e.g., minimum 100–200 events (42)] to avoid overfitting, particularly for rare outcomes like MDRO infections (28). (B) Handling missing data: Use advanced imputation methods [e.g., multiple imputation (35) or MissForest (36)] instead of complete-case analysis, which risks bias (28, 29). (C) Predictor selection: Combine clinical expertise (e.g., neonatology input for NICU models) with data-driven methods [e.g., LASSO regression, random forest variable importance (39, 40)] to avoid over-reliance on univariate analysis (2830). (2) Comprehensive validation: (A) External validation across diverse settings: Validate models in geographically distinct centers (e.g., comparing tertiary vs. community hospitals) and populations (e.g., NICUs vs. PICUs) to assess generalizability (34, 44). (B) Temporal validation: Include longitudinal cohorts to ensure stability of model performance over time, as demonstrated in two included studies (28, 29), but expanded to multi-year follow-up. (3) Advanced modeling techniques: Explore machine learning approaches [e.g., random forest, XGBoost (32, 33)] to capture non-linear relationships between predictors [e.g., interaction between antibiotic use and indwelling catheters (28, 30)] that traditional logistic regression may miss. Use ensemble methods to integrate the strengths of multiple models, potentially improving performance beyond single-algorithm approaches (33). (4) Enhanced transparency and clinical utility: (A) Strictly adhere to TRIPOD guidelines (19, 31), including full reporting of regression equations, calibration metrics, and validation protocols (absent in all included studies (2830). (B) Develop user-friendly tools (e.g., electronic health record-integrated calculators, mobile apps) and conduct usability testing with ICU nurses/physicians to ensure real-world applicability (47).

Strengths and limitations

This study is the first to conduct a comprehensive and integrated assessment of predictive models for NIs in ICU child patients. We provided valuable information for primary healthcare systems and clinical healthcare professionals through an extensive literature search, meticulous screening, and standardized data extraction. This approach lays the foundation for more effective construction and external validation of future predictive models. Furthermore, this study conducted a risk of bias and applicability assessment of the prediction models using the PROBAST tool, another significant strength.

This review is subject to certain limitations. Firstly, the literature search was confined to computerized databases and restricted to materials published in English and Chinese, potentially excluding relevant studies in other languages. Secondly, we could not perform meta-analyses or subgroup analyses due to the limited number of included studies and considerable heterogeneity in participant demographics, research settings, and outcomes. These variabilities could have affected the comparability and generalizability of the findings.

Conclusions

We identified three predictive models (one study was conducted in the PICU). In contrast, two studies were carried out in NICUs, and most of the researchers reported excellent discrimination and calibration in their research. However, for various reasons, the risk of bias in nearly all the models was high. Consequently, this finding implies that the predictive performance of these models might be overestimated, their accuracy in practical application to the target population remains questionable, and currently, we cannot endorse any of these predictive models for clinical practice. High-performance predictive models can assist clinical nursing staff in the early identification of high-risk populations for NIs, enabling proactive interventions to reduce infection rates. With the development of modeling methods, the selection of predictors largely determines the quality and effectiveness of the model. Predictors should be selected by combining clinical literature reports with data-driven approaches. Future research on predictive models for NI risk in ICU child patients should adhere to methodological guidelines, prioritize practicality and cost-effectiveness in model evaluation, conduct large-scale external validation, and ultimately facilitate effective identification of NIs in critically ill children.

Author contributions

LYW: Software, Conceptualization, Methodology, Writing – original draft, Investigation, Data curation. MF: Methodology, Writing – original draft. YLL: Writing – review & editing, Investigation. CXW: Writing – review & editing, Investigation. HW: Investigation, Funding acquisition, Writing – review & editing. LL: Investigation, Writing – review & editing. YZ: Data curation, Investigation, Writing – review & editing. XLH: Writing – review & editing, Data curation. MJH: Supervision, Investigation, Writing – review & editing. YMT: Methodology, Writing – review & editing, Conceptualization.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. The project supported by the Science and Technology Department of Sichuan Province (Project No. 2024JDKP0061) provided financial support for data collection and literature retrieval.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fped.2025.1636580/full#supplementary-material

References

1. Kollef MH, Torres A, Shorr AF, Martin-Loeches I, Micek ST. Nosocomial infection. Crit Care Med. (2021) 49:169–87. doi: 10.1097/CCM.0000000000004783

PubMed Abstract | Crossref Full Text | Google Scholar

2. Haque M, Sartelli M, McKimm J, Abu Bakar M. Health care-associated infections - an overview. Infect Drug Resist. (2018) 11:2321–33. doi: 10.2147/IDR.S177247

PubMed Abstract | Crossref Full Text | Google Scholar

3. Maki G, Zervos M. Health care-acquired infections in low- and middle-income countries and the role of infection prevention and control. Infect Dis Clin North Am. (2021) 35:827–39. doi: 10.1016/j.idc.2021.04.014

PubMed Abstract | Crossref Full Text | Google Scholar

4. Mazzeffi M, Galvagno S, Rock C. Prevention of healthcare-associated infections in intensive care unit patients. Anesthesiology. (2021) 135:1122–31. doi: 10.1097/ALN.0000000000004017

PubMed Abstract | Crossref Full Text | Google Scholar

5. Guzman-Cottrill JA, Bryant KA. Keeping your neonatal intensive care unit clean: the hospital environment as a potential source of health care-associated infections. Clin Perinatol. (2025) 52:1–14. doi: 10.1016/j.clp.2024.10.001

PubMed Abstract | Crossref Full Text | Google Scholar

6. Johnson J, Akinboyo IC, Schaffzin JK. Infection prevention in the neonatal intensive care unit. Clin Perinatol. (2021) 48:413–29. doi: 10.1016/j.clp.2021.03.011

PubMed Abstract | Crossref Full Text | Google Scholar

7. Donovan EF, Sparling K, Lake MR, Narendran V, Schibler K, Haberman B, et al. The investment case for preventing NICU-associated infections. Am J Perinatol. (2013) 30:179–84. doi: 10.1055/s-0032-1322516

PubMed Abstract | Crossref Full Text | Google Scholar

8. Patrick SW, Kawai AT, Kleinman K, Jin R, Vaz L, Gay C, et al. Health care-associated infections among critically ill children in the US, 2007–2012. Pediatrics. (2014) 134:705–12. doi: 10.1542/peds.2014-0613

PubMed Abstract | Crossref Full Text | Google Scholar

9. Hsu HE, Mathew R, Wang R, Broadwell C, Horan K, Jin R, et al. Health care-associated infections among critically ill children in the US, 2013–2018. JAMA Pediatr. (2020) 174:1176–83. doi: 10.1001/jamapediatrics.2020.3223

PubMed Abstract | Crossref Full Text | Google Scholar

10. Fitzgerald FC, Zingg W, Chimhini G, Chimhuya S, Wittmann S, Brotherton H, et al. The impact of interventions to prevent neonatal healthcare-associated infections in low- and middle-income countries: a systematic review. Pediatr Infect Dis J. (2022) 41:S26–35. doi: 10.1097/INF.0000000000003320

PubMed Abstract | Crossref Full Text | Google Scholar

11. Raoofi S, Pashazadeh Kan F, Rafiei S, Hosseinipalangi Z, Noorani Mejareh Z, Khani S, et al. Global prevalence of nosocomial infection: a systematic review and meta-analysis. PLoS One. (2023) 18:e0274248. doi: 10.1371/journal.pone.0274248

PubMed Abstract | Crossref Full Text | Google Scholar

12. Zhu S, Kang Y, Wang W, Cai L, Sun X, Zong Z. The clinical impacts and risk factors for non-central line-associated bloodstream infection in 5046 intensive care unit patients: an observational study based on electronic medical records. Crit Care. (2019) 23:52. doi: 10.1186/s13054-019-2353-5

PubMed Abstract | Crossref Full Text | Google Scholar

13. Nistal-Nuño B. A neural network for prediction of risk of nosocomial infection at intensive care units: a didactic preliminary model. Einstein (Sao Paulo). (2020) 18:eAO5480. doi: 10.31744/einstein_journal/2020AO5480

PubMed Abstract | Crossref Full Text | Google Scholar

14. Sharifzadeh Kermani M, Pouradeli S, Sadeghian R, Momen Abadi Z. Exploring the impact of comorbidities and drug resistance on mortality in ICU-acquired bloodstream infections. AMB Express. (2025) 15:70. doi: 10.1186/s13568-025-01874-w

PubMed Abstract | Crossref Full Text | Google Scholar

15. Pearl A, Bar-Or D. Using artificial neural networks to predict potential complications during trauma patients’ hospitalization period. Stud Health Technol Inform. (2009) 150:610–4. doi: 10.3233/978-1-60750-044-5-610

PubMed Abstract | Crossref Full Text | Google Scholar

16. Pearl A, Caspi R, Bar-Or D. Artificial neural network versus subjective scoring in predicting mortality in trauma patients. Stud Health Technol Inform. (2006) 124:1019–24.17108643

PubMed Abstract | Google Scholar

17. van Smeden M, Reitsma JB, Riley RD, Collins GS, Moons KG. Clinical prediction models: diagnosis versus prognosis. J Clin Epidemiol. (2021) 132:142–5. doi: 10.1016/j.jclinepi.2021.01.009

PubMed Abstract | Crossref Full Text | Google Scholar

18. Li J, Li G, Liu Z, Yang X, Yang Q. Prediction models for the risk of ventilator-associated pneumonia in patients on mechanical ventilation: a systematic review and meta-analysis. Am J Infect Control. (2024) 52:1438–51. doi: 10.1016/j.ajic.2024.07.006

PubMed Abstract | Crossref Full Text | Google Scholar

19. Collins GS, Reitsma JB, Altman DG, Moons KGM. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. Br Med J. (2015) 350:g7594. doi: 10.1136/bmj.g7594

PubMed Abstract | Crossref Full Text | Google Scholar

20. Ruppert MM, Loftus TJ, Small C, Li H, Ozrazgat-Baslanti T, Balch J, et al. Predictive modeling for readmission to intensive care: a systematic review. Crit Care Explor. (2023) 5:e0848. doi: 10.1097/CCE.0000000000000848

PubMed Abstract | Crossref Full Text | Google Scholar

21. Snell KIE, Levis B, Damen JAA, Dhiman P, Debray TPA, Hooft L, et al. Transparent reporting of multivariable prediction models for individual prognosis or diagnosis: checklist for systematic reviews and meta-analyses (TRIPOD-SRMA). Br Med J. (2023) 381:e073538. doi: 10.1136/bmj-2022-073538

PubMed Abstract | Crossref Full Text | Google Scholar

22. Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. Br Med J. (2021) 372:n71. doi: 10.1136/bmj.n71

PubMed Abstract | Crossref Full Text | Google Scholar

23. Li Y, Sperrin M, Ashcroft DM, van Staa TP. Consistency of variety of machine learning and statistical models in predicting clinical risks of individual patients: longitudinal cohort study using cardiovascular disease as exemplar. Br Med J. (2020) 371:m3919. doi: 10.1136/bmj.m3919

PubMed Abstract | Crossref Full Text | Google Scholar

24. Damen JAA, Moons KGM, van Smeden M, Hooft L. How to conduct a systematic review and meta-analysis of prognostic model studies. Clin Microbiol Infect. (2023) 29:434–40. doi: 10.1016/j.cmi.2022.07.019

PubMed Abstract | Crossref Full Text | Google Scholar

25. Alba AC, Agoritsas T, Walsh M, Hanna S, Iorio A, Devereaux PJ, et al. Discrimination and calibration of clinical prediction models: users’ guides to the medical literature. JAMA. (2017) 318:1377–84. doi: 10.1001/jama.2017.12126

PubMed Abstract | Crossref Full Text | Google Scholar

26. Blythe R, Parsons R, Barnett AG, Cook D, McPhail SM, White NM. Prioritising deteriorating patients using time-to-event analysis: prediction model development and internal-external validation. Crit Care. (2024) 28:247. doi: 10.1186/s13054-024-05021-y

PubMed Abstract | Crossref Full Text | Google Scholar

27. Moons KGM, Wolff RF, Riley RD, Whiting PF, Westwood M, Collins GS, et al. PROBAST: a tool to assess risk of bias and applicability of prediction model studies: explanation and elaboration. Ann Intern Med. (2019) 170:W1–33. doi: 10.7326/M18-1377

PubMed Abstract | Crossref Full Text | Google Scholar

28. Zhou J, Luo F, Liang J, Cheng X, Chen X, Li L, et al. Construction and validation of a predictive risk model for nosocomial infections with MDRO in NICUs: a multicenter observational study. Front Med (Lausanne). (2023) 10:1193935. doi: 10.3389/fmed.2023.1193935

PubMed Abstract | Crossref Full Text | Google Scholar

29. Miao YQ, Liu WW, Zhao SL, Jiang HM, Li YN, Teng P, et al. Construction and validation of risk prediction model for umbilical vein catheterization bloodstream infection. Chin J Infect Control. (2023) 22(02):159–66. (In Chinese). doi: 10.12138/j.issn.1671-9638.20233383

Crossref Full Text | Google Scholar

30. Saptharishi LG, Jayashree M, Singhi S. Development and validation of the “pediatric risk of nosocomial sepsis (PRiNS)” score for health care-associated infections in a medical pediatric intensive care unit of a developing economy–a prospective observational cohort study. J Crit Care. (2016) 32:152–8. doi: 10.1016/j.jcrc.2015.11.016

PubMed Abstract | Crossref Full Text | Google Scholar

31. Moons KG, Altman DG, Reitsma JB, Ioannidis JP, Macaskill P, Steyerberg EW, et al. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): explanation and elaboration. Ann Intern Med. (2015) 162:W1–73. doi: 10.7326/M14-0698

PubMed Abstract | Crossref Full Text | Google Scholar

32. Handelman GS, Kok HK, Chandra RV, Razavi AH, Lee MJ, Asadi H. Edoctor: machine learning and the future of medicine. J Intern Med. (2018) 284:603–19. doi: 10.1111/joim.12822

PubMed Abstract | Crossref Full Text | Google Scholar

33. Munger AM, Yu KE, Li DT, Furdock RJ, Boeyer ME, Duren DL, et al. The more the merrier: integrating multiple models of skeletal maturity improves the accuracy of growth prediction. J Pediatr Orthop. (2021) 41:e580–4. doi: 10.1097/BPO.0000000000001847

PubMed Abstract | Crossref Full Text | Google Scholar

34. Van Calster B, Steyerberg EW, Wynants L, van Smeden M. There is no such thing as a validated prediction model. BMC Med. (2023) 21:70. doi: 10.1186/s12916-023-02779-w

PubMed Abstract | Crossref Full Text | Google Scholar

35. Li Q, Yao X, Échevin D. How good is machine learning in predicting all-cause 30-day hospital readmission? Evidence from administrative data. Value Health. (2020) 23:1307–15. doi: 10.1016/j.jval.2020.06.009

PubMed Abstract | Crossref Full Text | Google Scholar

36. Zhou Z, Lin C, Ma J, Towne SD, Han Y, Fang Y. The association of social isolation with the risk of stroke among middle-aged and older adults in China. Am J Epidemiol. (2019) 188:1456–65. doi: 10.1093/aje/kwz099

PubMed Abstract | Crossref Full Text | Google Scholar

37. Zhou ZR, Wang WW, Li Y, Jin KR, Wang XY, Wang ZW, et al. In-depth mining of clinical data: the construction of clinical prediction model with R. Ann Transl Med. (2019) 7:796. doi: 10.21037/atm.2019.08.63

PubMed Abstract | Crossref Full Text | Google Scholar

38. Morozova O, Levina O, Uusküla A, Heimer R. Comparison of subset selection methods in linear regression in the context of health-related quality of life and substance abuse in Russia. BMC Med Res Methodol. (2015) 15:71. doi: 10.1186/s12874-015-0066-2

PubMed Abstract | Crossref Full Text | Google Scholar

39. Gu H, Liu C. Clinical prediction models: evaluation matters. Ann Transl Med. (2020) 8:72. doi: 10.21037/atm.2019.11.143

PubMed Abstract | Crossref Full Text | Google Scholar

40. Shen ZH. Research and Application of Dimensionality Reduction Techniques. Jiangnan University (2025). (In Chinese).

Google Scholar

41. Rosenthal VD, Yin R, Rodrigues C, Myatra SN, Divatia JV, Biswas SK, et al. Multinational prospective cohort study of incidence and risk factors for central line-associated bloodstream infections over 18years in 281 ICUs of 9 Asian countries. J Vasc Access. (2024) 25:1508–18. doi: 10.1177/11297298231169542

PubMed Abstract | Crossref Full Text | Google Scholar

42. Riley RD, Ensor J, Snell KIE, Harrell FE Jr, Martin GP, Reitsma JB, et al. Calculating the sample size required for developing a clinical prediction model. Br Med J. (2020) 368:m441. doi: 10.1136/bmj.m441

PubMed Abstract | Crossref Full Text | Google Scholar

43. Van Calster B, McLernon DJ, van Smeden M, Wynants L, Steyerberg EW. Calibration: the Achilles heel of predictive analytics. BMC Med. (2019) 17:230. doi: 10.1186/s12916-019-1466-7

PubMed Abstract | Crossref Full Text | Google Scholar

44. Grant SW, Collins GS, Nashef SAM. Statistical primer: developing and validating a risk prediction model. Eur J Cardiothorac Surg. (2018) 54:203–8. doi: 10.1093/ejcts/ezy180

PubMed Abstract | Crossref Full Text | Google Scholar

45. Xu Z, Li F, Xin Y, Wang Y, Wang Y. Prognostic risk prediction model for patients with acute exacerbation of chronic obstructive pulmonary disease (AECOPD): a systematic review and meta-analysis. Respir Res. (2024) 25:410. doi: 10.1186/s12931-024-03033-4

PubMed Abstract | Crossref Full Text | Google Scholar

46. Wynants L, Van Calster B, Collins GS, Riley RD, Heinze G, Schuit E, et al. Prediction models for diagnosis and prognosis of COVID-19: systematic review and critical appraisal. Br Med J. (2020) 369:m1328. doi: 10.1136/bmj.m1328

PubMed Abstract | Crossref Full Text | Google Scholar

47. Siette J, Dodds L, Sharifi F, Nguyen A, Baysari M, Seaman K, et al. Usability and acceptability of clinical dashboards in aged care: systematic review. JMIR Aging. (2023) 6:e42274. doi: 10.2196/42274

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: prediction, nosocomial infections, intensive care unit, child, model

Citation: Wang L-Y, Feng M, Luo Y-L, Wang C-X, Wang H, Li L, Zhang Y, Huang X-L, Huang M-J and Tian Y-M (2025) Predicting nosocomial infections in critically Ill children: a comprehensive systematic review of risk assessment models. Front. Pediatr. 13:1636580. doi: 10.3389/fped.2025.1636580

Received: 9 June 2025; Accepted: 18 August 2025;
Published: 10 September 2025.

Edited by:

Andrew Charles Argent, University of Cape Town, South Africa

Reviewed by:

Yaling Li, Peking University, China
Feng Zang, Pukou Branch of Jiangsu Province Hospital, China

Copyright: © 2025 Wang, Feng, Luo, Wang, Wang, Li, Zhang, Huang, Huang and Tian. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yong-Ming Tian, dGlhbnlvbmdtaW5nQHNjdS5lZHUuY24=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.