- 1Department of Hematology, Dongyang Hospital Affiliated with Wenzhou Medical University, Jinhua, China
- 2Department of Central Laboratory, Affiliated Hangzhou First People’s Hospital, School of Medicine, WestLake University, Hangzhou, China
Background: Radiographically confirmed pneumonia within 90 days of chemotherapy initiation is a frequent and clinically important complication in patients with non-Hodgkin lymphoma, yet interpretable tools for early individualized risk estimation are limited.
Objective: To develop and internally validate an interpretable machine-learning model that predicts the 90-day risk of radiographically confirmed pneumonia after chemotherapy initiation in non-Hodgkin lymphoma.
Methods: We retrospectively analyzed 205 chemotherapy-treated NHL patients. A two-step feature selection (LASSO followed by random-forest–based recursive feature elimination) identified four predictors: high-grade malignancy, drinking (alcohol use), estimated glomerular filtration rate (eGFR), and smoking. Five algorithms were trained and compared under a stratified 70/30 split (training n = 145; internal hold-out test set n = 60) with leakage-safe preprocessing (within-fold kNN imputation, SMOTE, and scaling). The gradient boosting machine (GBM) performed best and was interpreted using SHAP. A web-based prototype was implemented for research use only.
Results: On the internal hold-out test set (n = 60), the GBM achieved an AUC of 0.855 (95% CI 0.746–0.964), an F1 score of 0.679, and a Brier score of 0.155. SHAP identified reduced eGFR, smoking, drinking, and high-grade malignancy as influential contributors; case-level waterfall and force plots enhanced transparency. These estimates reflect internal validation only and were obtained without systematic microbiological confirmation or standardized radiologic rescoring. Accordingly, performance may be optimistic, and real-world use is not advised pending temporal and multicenter external validation (with potential recalibration) and prospective evaluation.
Conclusion: The interpretable GBM model demonstrated promising discrimination and calibration on an internal hold-out test set; however, clinical deployment requires temporal and multicenter external validation (as well as prospective assessment with potential recalibration). The accompanying web calculator is a research-only prototype and is not intended for clinical decision-making until such validation is completed.
1 Introduction
Non-Hodgkin lymphoma (NHL) is a heterogeneous group of lymphoid malignancies for which global incidence continues to rise, reaching an estimated 545,000 new cases and 260,000 deaths in 2020 (1, 2). The widespread use of immunochemotherapy regimens, such as R-CHOP, along with the introduction of targeted therapies, has significantly improved survival. However, infectious complications—particularly pneumonia—remain among the most common and clinically significant adverse events during chemotherapy. In elderly NHL patients treated with R-CHOP, pulmonary complications have been reported in up to 40 percent of cases, with approximately 10 percent experiencing severe infections (3). A previous study involving 229 newly diagnosed NHL patients reported that 91 (39.7%) developed bacterial pneumonia, including 76 with isolated respiratory tract infection and 15 with concurrent bacteremia (4). Despite its frequency, pneumonia is often underrecognized in the early stages due to insidious and nonspecific symptoms, which can delay diagnosis and treatment, thereby adversely affecting clinical outcomes (5, 6). Although pneumonia is clinically burdensome—especially during the first 90 days after chemotherapy initiation, when patients are most vulnerable, validated tools for early risk prediction in this critical period remain limited. Few studies have developed individualized prediction models tailored to this early and high-risk phase of treatment.
Pneumonia in patients with NHL most commonly occurs within the first 90 days after chemotherapy initiation. This period is typically marked by bone marrow suppression, disruption of mucosal barriers, and compromised immune function (6–8). Early identification of high-risk individuals during this timeframe is essential to guide timely preventive interventions, such as empiric antimicrobial therapy, antifungal prophylaxis, or chemotherapy dose adjustments. However, existing risk stratification tools, including the MASCC score, were not specifically designed to predict pneumonia in patients with NHL, highlighting a critical gap in early risk assessment (9).
Several predictive tools exist in oncology for infection-related adverse outcomes, but they target different questions than organ-specific pneumonia risk. Febrile neutropenia (FN) risk stratification models—such as the Multinational Association for Supportive Care in Cancer (MASCC) Risk Index, Talcott classification, and the Clinical Index of Stable Febrile Neutropenia (CISNE)—were designed to identify low- versus high-risk FN episodes and guide outpatient management, not to estimate a patient’s individualized probability of developing pneumonia in the early chemotherapy period (10–13). In parallel, risk scores for invasive fungal disease (e.g., invasive mold disease) have been proposed primarily for hematologic malignancies or transplant settings, but these focus on a fungal subset rather than the broader spectrum of bacterial, viral, and fungal pneumonia encountered during NHL chemotherapy (14, 15). Consequently, current tools leave a gap for early, organ-specific prediction of pneumonia risk in NHL.
Pneumonia warrants prioritization for several clinically relevant reasons. First, the lung is a common infection site in patients with hematologic malignancies receiving chemotherapy; pneumonia contributes disproportionately to ICU admissions and mortality compared with many other infection sites (16, 17). Second, early detection is actionable: timely imaging, empiric antimicrobial therapy, selective antifungal prophylaxis, vaccination strategies, and—when appropriate—chemotherapy dose adjustments can be targeted to those at highest risk (18). Third, pneumonia’s insidious early presentation in immunocompromised hosts underscores the need for individualized, pre-emptive risk estimation to avoid delays in diagnosis and treatment.
To address this need, we developed and validated a machine-learning-based model to predict radiographically confirmed pneumonia occurring within 90 days after chemotherapy initiation in patients with NHL. Key predictors were identified using random forest–based recursive feature elimination (RF-RFE), and five machine learning algorithms were constructed and compared. Among these, the gradient boosting machine (GBM) model demonstrated superior predictive performance. To enhance interpretability and clinical applicability, SHAP (Shapley Additive Explanations) was applied to explain model predictions at the individual level, and the final model was implemented as an interactive web-based tool (19). This study provides a practical and individualized solution for early risk assessment of pneumonia in patients with NHL undergoing chemotherapy.
2 Methods
2.1 Study design and data source
2.1.1 Study setting and cohort
This retrospective study was conducted at Dongyang Hospital Affiliated with Wenzhou Medical University included consecutive patients with pathologically confirmed non-Hodgkin lymphoma (NHL) who initiated systemic chemotherapy between October 2018 and October 2024. Eligibility required receipt of at least one chemotherapy cycle and follow-up for at least 90 days after initiation or until the earlier occurrence of pneumonia or death. We excluded patients with radiographically confirmed pneumonia prior to the initiation of chemotherapy and those with incomplete baseline or follow-up data.
2.1.2 Primary outcome
The primary outcome was radiographically confirmed pneumonia occurring within 90 days of chemotherapy initiation. Case ascertainment required (i) a new or progressive pulmonary infiltrate on chest radiography or computed tomography and (ii) ≥ 1 clinical criterion: new or worsened cough or sputum; fever; auscultatory findings consistent with consolidation (e.g., crackles/wet rales, bronchial breath sounds, or egophony); or an abnormal peripheral leukocyte count > 10 × 109/L or < 4 × 109/L (with or without left shift). Episodes without a new infiltrate (e.g., tracheobronchitis/URTI, COPD/asthma exacerbation) and non-infectious mimics (drug- or radiation-induced pneumonitis, cardiogenic pulmonary edema, diffuse alveolar hemorrhage, pulmonary embolism/infarction, tumor progression/lymphangitic carcinomatosis) were excluded. Both community-onset and hospital-onset episodes meeting this definition were eligible. Because the 2019 ATS/IDSA community-acquired pneumonia guideline excludes immunocompromised hosts, we adopted its radiographic–clinical framework for case ascertainment only and did not apply CAP-specific care pathways.
2.1.3 Data availability for phenotyping
Microbiological data (e.g., culture/PCR panels) were not systematically collected, and imaging studies were not centrally reviewed or scored using a standardized severity metric; case ascertainment relied on clinical radiology reports according to the operational definition.
2.2 Variable collection and preprocessing
A total of 35 clinical variables were collected, covering demographic characteristics, disease status, comorbidities, treatment-related factors, and baseline laboratory results. All predictors were pre-specified and dichotomized to binary indicators (0/1) before modeling; no multilevel factors remained (20). Missing values were imputed using k-nearest neighbors (kNN) with k = 5; imputation models were fit on the training data (or within cross-validation folds) and then applied to held-out data only to avoid information leakage (21). Because kNN imputation can yield fractional values for originally binary predictors, those fields were post-processed using a 0.5 threshold to restore 0/1 coding (22). Robustness to the imputation hyperparameter was examined in an alternative specification by varying k (3 and 7). Since this specification differs from the primary pipeline (settings and implementation), absolute performance values are reported in Supplementary Table S8 and are not intended for direct comparison with the primary test-set results; qualitative conclusions were unchanged.
2.3 Statistical analysis (baseline comparisons)
Baseline characteristics were summarized as n (%) for categorical variables. For comparisons between the pneumonia and no-pneumonia groups (Table 1), group differences were assessed using two-sided tests (α = 0.05): Pearson’s chi-square test with Yates’ continuity correction, or Fisher’s exact test when any expected cell count was < 5. Missing values were excluded from hypothesis testing. All baseline variables were dichotomous for presentation; binary 0/1 coding was used for model development (Section 2.2).

Table 1. Baseline characteristics of patients with non-Hodgkin lymphoma (NHL) with and without pneumonia within 90 days after chemotherapy initiation.
For comparisons between the training and test sets (Table 2), standardized mean differences (SMDs) were reported as effect sizes that are less sensitive to sample size (imbalance flagged at |SMD| > 0.20; |SMD| < 0.10 generally negligible) (23–25). Train–test comparability was summarized with SMDs only, without p values. A pre-specified sensitivity analysis excluding variables with |SMD| > 0.20 yielded essentially unchanged results (Supplementary Tables S9, S10).

Table 2. Baseline characteristics of the training set and the internal hold-out test set (derived by stratified random split from the same single-center cohort).
Analyses were performed in R (version 4.4.2) using the TableOne package (CreateTableOne) and base R functions. Methods for AUC (DeLong), bootstrap 95% CIs for threshold-based metrics, calibration, and decision-curve analysis are detailed in Section 2.7.
2.4 Data split and internal validation
We performed a stratified 70/30 split based on the outcome (pneumonia vs. no pneumonia) to create a training set (n = 145) and an internal hold-out test set (n = 60) from the same single-center cohort, ensuring balanced class distributions across sets. All preprocessing steps—including kNN imputation, feature selection (LASSO followed by RF-RFE), class-imbalance handling (SMOTE within cross-validation), and hyperparameter tuning—were conducted strictly on the training data. These procedures were confined to cross-validation folds to prevent information leakage. The hold-out test set was left untouched until final evaluation. Because the internal hold-out test set was modest (n = 60; events = 23), we treated performance from repeated 10-fold cross-validation in the training set as the primary estimate, while test-set metrics served as complementary internal validation and are reported with 95% CIs (AUC via DeLong; other metrics via class-stratified bootstrap, B = 2,000).
Baseline comparability between the training and test sets was summarized using standardized mean differences (SMDs) and is presented in Table 2. We did not report hypothesis-test p values for the train-test comparison because such tests are highly sample-size dependent and less informative for balance assessment in a small internal test set. Importantly, the internal hold-out test set was never resampled and was kept completely isolated from all preprocessing, imputation, SMOTE, and feature-selection procedures (see Figure 1).

Figure 1. Flowchart of patient selection and machine-learning workflow for predicting radiographically confirmed pneumonia within 90 days after chemotherapy initiation in NHL patients.
2.5 Feature selection
To identify a parsimonious and stable set of predictors, we used a two-step procedure. First, least absolute shrinkage and selection operator (LASSO) logistic regression was applied to the full set of 35 candidate variables with 10-fold cross-validation; to favor sparsity and reduce variance, we selected the λ.1se solution (26). In our data, LASSO (λ.1se) retained four variables—high-grade malignancy, drinking, eGFR, and smoking. Second, we performed events-per-variable (EPV)—constrained random-forest–based recursive feature elimination (RF-RFE) on the LASSO-retained variables, enforcing EPV ≥ 5 and a maximum of 10 predictors, which was appropriate for the 79 pneumonia events (27). We prespecified a stopping rule to select the smallest subset on the performance plateau, defined as ∆AUC ≤ 0.01 from the maximal cross-validated AUC. When adjacent subsets both satisfied the plateau criterion, a pre-specified tie-breaker favored the subset with the higher mean CV AUC and greater selection stability; this yielded the final four-predictor model. The incremental AUC-by-subset curve is shown in (Figure 2C), and stepwise subsets are tabulated in (Supplementary Table S2). To assess robustness, we conducted nested cross-validation (outer 5-fold, 20 repeats): within each outer training fold, LASSO (λ.1se) and EPV-constrained RFE were refit, and performance was evaluated on the corresponding outer test fold; selection frequencies across repeats were summarized (Supplementary Figure 1A; Supplementary Table S3). In addition, we performed 200 bootstrap resamples of the LASSO step (λ.1se) to estimate per-variable selection probabilities (Supplementary Table S4). Potential multicollinearity among final predictors was screened using Pearson correlations (after dummy expansion) and variance inflation factors (VIF/GVIF); we flagged |r| > 0.70 or VIF > 5 (GVIF_adj > 2) as concerning. The correlation heatmap appears in (Figure 2D), with numeric diagnostics in (Supplementary Table S5).

Figure 2. Two-stage feature selection for predicting radiographically confirmed pneumonia within 90 days after chemotherapy initiation in NHL patients. (A) LASSO coefficient profiles across Log Lambda; vertical dashed lines mark Lambda min (orange) and Lambda 1se (blue). Numbers along the top axis indicate the count of non-zero coefficients at each Lambda. (B) LASSO 10-fold cross-validation curve showing mean binomial deviance (red) with SE bars versus Log Lambda; the parsimonious Lambda 1se was adopted for variable retention. (C) Cross-validated AUC by subset size during EPV-constrained RF-RFE after LASSO selection; the optimal subset comprised four predictors. (D) Correlation among the final selected predictors (eGFR, smoking, drinking, high-grade malignancy). numeric collinearity diagnostics (VIF/GVIF) are provided in Supplementary Table S5. Feature selection and tuning were performed strictly within training cross-validation folds; the internal hold-out test set was not used for selection.
2.6 Hyperparameter tuning and model selection
Hyperparameter tuning was performed exclusively on the training set to avoid information leakage. For the support vector machine (SVM), GBM, and k-nearest neighbors (KNN), we employed Bayesian optimization (rBayesianOptimization, upper confidence bound acquisition with κ = 2.0). Each candidate hyperparameter configuration was evaluated using 5-fold cross-validation, with SMOTE applied strictly within folds. The binary outcome was coded as pneumonia = “Yes” and no pneumonia = “No”; model selection optimized the mean cross-validated AUC (caret twoClassSummary, positive = “Yes”). For SVM and KNN only, predictors were z-score centered and scaled during both the 5-fold tuning stage and the subsequent repeated 10-fold cross-validation stage used for performance estimation and final refitting; scaling parameters were estimated on the in-fold training (analysis) partition and applied to the paired validation (assessment) partition, with no carryover across folds. After model selection, the centering/scaling transform was refit on the full training set and applied once to the internal hold-out test set before final evaluation. Other models (logistic regression, GBM, LightGBM) did not use feature scaling. Logistic regression did not require hyperparameter tuning.
For the light gradient boosting machine (LightGBM), a two-stage procedure was adopted. First, Bayesian search with 4-fold cross-validation and early stopping was performed across predefined ranges for learning rate, num_leaves, min_data_in_leaf, feature_fraction, bagging_fraction, λ1, and λ2. Second, the optimal parameter set underwent confirmatory 5-fold cross-validation to determine the best iteration number, after which the final model was refitted on the full training set for nrounds = best_iter. The internal hold-out test set was never used for resampling, tuning, or threshold determination. Detailed search bounds and final selected hyperparameter values for all models are provided in Supplementary Table S6.
2.7 Model development and comparison
To mitigate class imbalance during model development, we applied SMOTE within each resampling fold of cross-validation using caret:trainControl (sampling = “smote”). The same within-fold procedure was used during hyperparameter tuning. No resampling was ever applied to the internal hold-out test set. By design, this prevents information leakage and yields unbiased validation estimates (28). Five machine-learning models were constructed based on the selected predictors: logistic regression, SVM, KNN, GBM, and LightGBM. Hyperparameters were optimized via Bayesian optimization evaluated with 5-fold cross-validation on the training set (for LightGBM: 4-fold during the search with early stopping plus confirmatory 5-fold to determine best_iter), as detailed in Section 2.6. Repeated 10-fold cross-validation (10 × 5) on the training set was then used for performance estimation and to generate out-of-fold predictions; the internal hold-out test set was used once for final evaluation. Model performance was evaluated on the internal hold-out test set using multiple metrics, including the AUC, F1 score, Brier score, accuracy, sensitivity, specificity, and predictive values. Decision thresholds were pre-specified on the training set by maximizing Youden’s J and then fixed for both training and test evaluations. We report AUC with DeLong 95% confidence intervals (CIs); at the fixed threshold we report accuracy, sensitivity, specificity, PPV, NPV, and F1 score with class-stratified bootstrap 95% CIs (B = 2,000), together with confusion-matrix counts (TP, TN, FP, FN). Calibration was assessed using calibration curves and the Brier score with bootstrap 95% CIs, and clinical utility was evaluated using decision-curve analysis (DCA) on predicted probabilities. All preprocessing resampling (SMOTE), hyperparameter tuning, and threshold selection were confined to training-set cross-validation; the test set was used only once for final evaluation.
2.8 Model interpretation
To enhance the interpretability of the final GBM model, a post hoc analysis was conducted using SHAP, a game-theoretic approach that quantifies the contribution of individual predictors to the model outputs. Global feature importance was visualized using a SHAP summary bar plot and a beeswarm plot, which illustrate the relative magnitude and direction of each variable’s impact on the predicted risk of pneumonia. For individual-level interpretation, a SHAP waterfall plot and a force plot were generated to demonstrate how specific predictor values influenced the deviation from the model’s baseline prediction, thereby providing transparent, case-specific insights into the model’s decision-making process (29, 30).
2.9 Web deployment
The final GBM model was implemented as an interactive R Shiny web application for research use. This platform accepts patient-specific clinical data and computes risk estimates with explanations for research evaluation. The tool is a prototype for internal validation and has not undergone prospective clinical implementation or impact evaluation; it is not intended to guide patient care or clinical decision-making and has not been integrated into routine clinical workflows (31, 32).
3 Results
3.1 Baseline characteristics
A total of 205 patients diagnosed with NHL were included in the final analysis, among whom 79 (38.5%) developed pneumonia within 90 days after chemotherapy initiation. Baseline characteristics were compared between the pneumonia group (n = 79) and the no-pneumonia group (n = 126) (Table 1). Patients with pneumonia were more often male (69.6% vs. 48.4%, p = 0.005) and more frequently had advanced Ann Arbor stage III–IV disease (78.5% vs. 58.7%, p = 0.006). Significant differences were also observed in smoking (57.0% vs. 16.7%, p < 0.001), drinking (58.2% vs. 17.5%, p < 0.001), high-grade malignancy (81.0% vs. 46.0%, p < 0.001), and reduced renal function (eGFR < 80 mL/min/1.73 m2, 51.9% vs. 9.5%, p < 0.001). In contrast, there were no statistically significant differences between groups for age (p = 0.189), NCCN-IPI (p = 0.219), post-chemotherapy marrow suppression (p = 0.702), hepatitis B status (p = 0.246), BMI (p = 0.788), hypertension (p = 1.000), diabetes (p = 0.624), treatment regimen (p = 0.290), or the majority of baseline hematologic/biochemical indices (all p > 0.05), including WBC, Hb, ANC, ALC, RDW, PLT, MPV, PCT, calcium, iron, albumin, A/G ratio, ALT, AST, LDH, α-HBDH, triglycerides, HDL-C, LDL-C, and CRP.
For model development, we then performed a stratified 70/30 split by the outcome to create a training set (n = 145) and an internal hold-out test set (n = 60) from the same single-center cohort. Between-set comparability was summarized using standardized mean differences (SMDs) (Table 2). Most characteristics were well balanced (|SMD| < 0.20). Four laboratory variables—WBC, ANC, PLT, and CRP—showed modest imbalance (|SMD| ≈ 0.21, 0.21, 0.21, and 0.39, respectively). A prespecified sensitivity analysis excluding these variables is described in Section 3.4 and Supplementary Tables S9, S10.
3.2 Feature selection results
Using the pre-specified two-step procedure, LASSO (λ.1se) retained four variables—high-grade malignancy, drinking, eGFR, and smoking (Figures 2A,B). Applying events-per-variable–constrained RF-RFE to this set, the cross-validated AUC increased with the number of features and reached a performance plateau at 3–4 predictors (Figure 2C). In our data, the mean cross-validation (CV) AUCs were 0.803 for k = 3 and 0.804 for k = 4 (∆AUC = 0.001, within the pre-specified plateau tolerance ∆AUC ≤ 0.01; Supplementary Table S2). Per our tie-breaker for adjacent subsets on the plateau—which selected the subset with the higher mean CV AUC and greater selection stability—we retained the four-predictor model (high-grade malignancy, drinking, eGFR, smoking) for subsequent development and interpretation. Stability was high in nested cross-validation (outer 5-fold, 20 repeats), with consistently elevated selection frequencies for these four predictors (Supplementary Table S3; Supplementary Figure S1A). Bootstrap resampling of the LASSO step (B = 200) produced concordant stability results (Supplementary Table S4). Collinearity diagnostics (VIF/GVIF and Pearson correlations) indicated no evidence of concerning multicollinearity among the final predictors (Figure 2D; Supplementary Table S5).
3.3 Model performance comparison
Model performance was reported at a pre-specified threshold fixed from the training set by Youden’s J. All threshold-based metrics were presented with class-stratified bootstrap 95% CIs; AUC 95% CIs were DeLong-based. Five machine-learning models—logistic regression, SVM, KNN, GBM, and LightGBM—were all trained using the same four selected predictors, avoiding model-specific reselection and potential selection-induced bias. The performance of each model was subsequently evaluated on both the training set and an internal hold-out test set (n = 60).
On the training set, GBM achieved the highest AUC (0.853, 95% CI 0.789–0.916), followed by SVM (0.844, 95% CI 0.778–0.910), LightGBM (0.843, 95% CI 0.777–0.909), logistic regression (0.841, 95% CI 0.775–0.908), and KNN (0.806, 95% CI 0.739–0.874) (Figure 3A). On the internal hold-out test set, GBM achieved the highest AUC (0.855, 95% CI 0.746–0.964), followed by logistic regression (0.844, 95% CI 0.732–0.957), LightGBM and SVM (both 0.841, 95% CI 0.729–0.953), and KNN (0.588, 95% CI 0.451–0.724) (Figure 3B). These results indicate that GBM provided the best overall discrimination in both training cross-validation and internal testing.

Figure 3. Comparative performance of five machine-learning models for predicting radiographically confirmed pneumonia within 90 days after chemotherapy initiation in NHL patients. (A,B) ROC curves for the training set (A) and the internal hold-out test set (B); AUCs with DeLong 95% CIs are reported in the legends. (C,D) Radar plots summarizing threshold-based metrics (AUC, accuracy, sensitivity, specificity, PPV, F1) for the training set (C) and the internal hold-out test set (D) at the pre-specified threshold (Youden’s J, determined from the training set and fixed for testing); corresponding 95% CIs are provided in Table 3. (E,F) Calibration curves with corresponding Brier scores (bootstrap 95% CIs) for the training set (E) and the internal hold-out test set (F). (G,H) Decision-curve analysis in the training set (G) and the internal hold-out test set (H), showing net benefit across threshold probabilities; “all” and “none” strategies are included for reference. (I,J) Confusion matrices of the GBM model for the training set (I) and the internal hold-out test set (J; n = 60) at the pre-specified threshold. ROC CIs are DeLong-based; threshold-based metrics and Brier CIs use class-stratified bootstrap. All preprocessing, feature selection, SMOTE, and tuning were confined to training cross-validation; the internal test set was not used for resampling or threshold selection.
Full threshold-based metrics with 95% CIs and the corresponding confusion-matrix counts (TP/TN/FP/FN) for all models are provided in Table 3. For completeness, training-set metrics with 95% CIs are provided in Supplementary Table S7. As illustrated in the radar plots (Figures 3C–D), GBM outperformed the other algorithms in terms of overall discrimination (AUC) and achieved balanced sensitivity and predictive values. LightGBM yielded slightly higher accuracy (0.750 vs. 0.717) and specificity (0.730 vs. 0.676), but GBM maintained the most robust overall performance across metrics. Calibration curves demonstrated good agreement between predicted and observed risks for GBM, with Brier scores of 0.151 (training) and 0.155 (internal test) (Figures 3E–F; Table 3). Decision-curve analysis showed comparable net benefit for GBM, logistic regression, SVM, and LightGBM across most clinically relevant thresholds, with no uniform winner; KNN underperformed (Figures 3G–H). These patterns support decision-analytic utility under internal validation. The confusion matrices further illustrate the classification performance of the GBM model in both the training set (Figure 3I) and the internal hold-out test set (Figure 3J). In the training set, correct predictions predominated, with relatively balanced error rates across positive and negative classes. In the test set, although overall discrimination was maintained, a higher proportion of false positives was observed, reflecting a modest decrease in specificity.

Table 3. Performance of machine-learning models on the internal hold-out test set using pre-specified training-set thresholds.
Collectively, these findings support GBM as the most robust model for individualized pneumonia-risk prediction under internal validation. All estimates reflect internal validation only; in the absence of routine microbiological confirmation, standardized radiologic scoring, and any real-world/prospective validation, performance may be optimistic and not ready for clinical deployment.
3.4 Sensitivity analysis of imputation
To assess robustness to imputation, we examined an alternative specification that varied the k parameter in kNN imputation (k = 3, 5, and 7). Because this specification differed from the primary pipeline, absolute performance values were presented in Supplementary Table S8 and should not be interpreted as head-to-head comparisons with the primary test-set results. Results were qualitatively unchanged. Second, to address modest baseline imbalances, we repeated the primary pipeline without changes (i.e., same preprocessing, cross-validation/tuning protocol, and model class) but excluded variables with |SMD| > 0.20 between the training and test sets (WBC, ANC, PLT, CRP). Performance was essentially unchanged relative to the primary analysis (Supplementary Tables S9, S10), indicating that these variables did not materially influence model performance.
3.5 SHAP-based model interpretation
To improve the interpretability of the GBM model, SHAP was applied to quantify the contribution of each predictor to the model’s output. The SHAP summary bar plot (Figure 4A) ranked the four selected features by their mean absolute SHAP values, with eGFR showing the strongest overall influence, followed by smoking, drinking, and high-grade malignancy. The SHAP beeswarm plot (Figure 4B) visualized the distribution and direction of each feature’s contribution. Generally, higher SHAP values for high-grade malignancy and lower eGFR were associated with increased predicted risk of pneumonia, while the effects of smoking and drinking varied across individuals.

Figure 4. SHAP-based interpretation of the GBM model for predicting radiographically confirmed pneumonia within 90 days after chemotherapy initiation in NHL patients. (A) Global feature importance ranked by mean absolute SHAP value. (B) SHAP beeswarm plot showing the distribution, direction, and magnitude of feature contributions across the internal hold-out test set (points colored by feature value). (C) SHAP waterfall plot for a representative high-risk case (Smoking = 1, Drinking = 1, eGFR = 1, High-grade malignancy = 0), illustrating how individual feature contributions shift the predicted probability from the baseline E[f(x)] = 0.519 to f(x) = 0.988. (D) SHAP force plot for the same case, summarizing how these contributions combine to yield the final prediction. Predictor encoding: Smoking (1 = Yes, 0 = No), Drinking (1 = Yes, 0 = No), High-grade malignancy (1 = Yes, 0 = No), eGFR (1 = < 80 mL/min/1.73 m2, 0 = ≥ 80 mL/min/1.73 m2).
For case-level interpretation, waterfall and force plots (Figures 4C,D) were generated for a representative high-risk patient. In this example, concurrent smoking, drinking, and reduced renal function substantially elevated the model’s predicted risk despite the absence of high-grade malignancy, as reflected by a high model output (f(x) = 0.988 vs. E[f(x)] = 0.519). To complement this, an illustrative low-risk case was presented in Supplementary Figures S1B,C, where the absence of high-grade malignancy, preserved renal function (eGFR = 0), and lack of smoking and drinking collectively reduced the predicted probability well below the baseline expectation (f(x) = 0.186 vs. E[f(x)] = 0.519).
These individualized explanations improve transparency, highlight clinically actionable factors, and provide practical insights for prevention. For example, smoking cessation and alcohol avoidance represent modifiable targets, while vigilant monitoring of renal function and high-grade disease may guide early intervention. Together, these case-level insights support patient-centered decision-making and strengthen the clinical applicability of the model. Such individualized visualizations enhance model transparency and reinforce personalized clinical decision-making.
3.6 Web-based prediction tool deployment
To improve clinical applicability, the final GBM model was implemented as an interactive web-based tool using the R Shiny framework. This online platform allows healthcare providers to input patient-specific clinical parameters and obtain real-time estimates of the probability of pneumonia occurring within 90 days of chemotherapy initiation. The web-based tool interface contains four input fields, namely high-grade malignancy, drinking, eGFR, and smoking (Figure 5). After data entry, the system promptly computes and displays the predicted probability of pneumonia, providing a quantitative reference to support clinical evaluation. The tool is accessible through standard web browsers on both desktop and mobile devices, without the need for additional software installation. Its real-time output and integration of SHAP-based interpretive visualizations enhance transparency and support individualized risk assessment in routine hematology practice. The tool is available at: https://ltfu-zzn.shinyapps.io/Pneumonia/

Figure 5. Prototype web-based calculator for individualized pneumonia risk prediction in NHL patients after chemotherapy initiation. The R Shiny interface allows entry of four predictors (smoking, drinking, high-grade malignancy, eGFR) to generate an individualized 90-day pneumonia risk score with interpretive visualization.
4 Discussion
4.1 Principal findings
In this study, we developed a machine learning-based model to predict early pneumonia in patients with NHL undergoing chemotherapy. The model was constructed using four clinically relevant predictors: high-grade malignancy, drinking status, eGFR, and smoking status. Among the five algorithms evaluated, the GBM model achieved the highest predictive performance. On the internal hold-out test set, it yielded an AUC of 0.855, an F1 score of 0.679, and a Brier score of 0.155, demonstrating favorable discrimination, calibration, and overall predictive accuracy. Among the evaluated algorithms, GBM achieved a favorable sensitivity–specificity balance and maintained consistent performance across the training set and the internal hold-out test set. The use of SHAP values further facilitated the interpretation of model predictions at both the global and individual levels. Impaired renal function, and behavioral factors such as smoking and drinking were identified as key contributors to pneumonia risk. These findings support the utility of interpretable machine learning techniques in predicting early pneumonia risk in NHL patients, offering a potential tool for personalized risk stratification and preventive care in hematologic oncology. Because our evaluation used an internal hold-out split from the same single-center cohort, true external generalizability remains untested. We therefore plan to conduct temporal validation in a later patient cohort and pursue multicenter external validation to assess transportability; if dataset shift is observed, model recalibration and threshold re-specification will be performed.
4.2 Clinical and scientific implications
A previous study reported that 39.7% of newly diagnosed NHL patients developed bacterial pulmonary infections (4). After chemotherapy, the risk of such infections in NHL patients increases significantly. Other studies have shown that most severe infections occur within six months of initiating rituximab-based chemotherapy (33), emphasizing the need for early and continuous monitoring. Such infections frequently lead to treatment delays or discontinuation and were associated with substantially poorer progression-free survival. Accurate assessment of infection risk in the early treatment phase therefore enables timely implementation of preventive measures, including prophylactic antibiotics, enhanced monitoring, and tailored chemotherapy regimens.
Beyond positioning our work relative to MASCC, prior interpretable ML studies in related populations help contextualize our contribution. In lung cancer, Sun et al. predicted post-chemotherapy lung infection using 36 clinical/laboratory predictors selected by Boruta/LASSO; a regularized logistic-regression model with SHAP achieved an AUC ≈ of 0.89 and reported calibration and decision-curve analyses. Their endpoint required clinical signs plus a CT-identified infectious focus, but was not restricted to radiographically confirmed pneumonia within a prespecified 90-day window (34). In newly diagnosed multiple myeloma, Peng et al. trained models on 38 variables and found XGBoost performed best (AUC ≈ 0.88 in validation), again with SHAP explanations; the outcome was a composite infection definition (microbiologically or clinically defined infection, or FUO) rather than pneumonia specifically (35). Both studies emphasized discrimination and interpretability using internal resampling for validation. By contrast, our study focuses on a narrow, clinically actionable endpoint—radiographically confirmed pneumonia within 90 days of chemotherapy initiation in NHL—and demonstrates competitive performance through a parsimonious, interpretable gradient boosting machine model explained by SHAP, incorporating four predictors (high-grade malignancy, eGFR, smoking, and drinking).
This focus is supported by real-world data indicating that nearly half of respiratory complications in NHL occur during this early treatment window, with approximately 75% being infectious in origin (6). The median onset of interstitial pneumonia at 74 days after treatment initiation further illustrates the vulnerability of this phase (36), even though interstitial pneumonia itself is non-infectious. In line with previous studies, smoking was reaffirmed as a risk factor for pneumonia, while high-grade malignancy also emerged as a predictor, plausibly reflecting immune dysfunction in aggressive subtypes such as DLBCL (37). Retrospective analyses of DLBCL have shown high rates of infectious episodes (38) and substantially reduced five-year survival in patients who developed pneumonia during chemotherapy (41% vs. 82%) (39). Similarly, reduced eGFR was associated with pneumonia risk, underscoring the contribution of renal dysfunction during immunochemotherapy (40). Alcohol use was another significant predictor, consistent with meta-analytic evidence linking alcohol consumption to impaired pulmonary immune defenses (41).
Taken together, this study represents a novel application of interpretable ML for pneumonia risk management in NHL. By leveraging GBM with SHAP explanations and decision-curve analysis, we provide not only a high-performing predictive model but also a transparent, clinician-friendly tool. While SHAP offers individualized, transparent rationale for risk estimates, it reflects associations rather than causation; therefore, explanations should be used to inform preventive vigilance and shared decision-making, with thresholds and calibration subject to confirmation in external/temporal validation. Unlike conventional scoring systems, our model captures non-linear relationships and interactions among malignancy severity, disease stage, renal function, and behavioral factors, thereby offering a practical and data-driven approach to pneumonia surveillance in hematologic malignancies.
4.3 Comparison with existing risk models
Several clinical scoring tools have been used to assess infection risk in cancer patients, with the MASCC score being one of the most commonly used. However, it was originally developed for febrile neutropenia and may not be well suited to predicting pneumonia, especially in lymphoma patients. The score is based on general clinical parameters, such as the burden of illness and outpatient status, but lacks consideration of tumor biology or treatment-specific factors (42, 43). Our model, in contrast, was developed specifically for NHL patients receiving chemotherapy, with a focus on predicting pneumonia during the early, high-risk phase of treatment. By including variables such as eGFR and high-grade malignancy, it achieved strong performance (AUC = 0.855 on the internal hold-out test set) and may offer advantages over conventional additive scores. Unlike conventional tools that rely on additive point systems and assume linear effects, our GBM model can account for complex interactions among clinical features. Importantly, the use of SHAP values allows the model to explain individual predictions—something traditional scores do not provide (44). Overall, our work offers a more focused and clinically adaptable approach to pneumonia risk assessment in hematologic oncology, complementing existing models while addressing some of their key limitations.
4.4 Limitations
This study has several limitations that should be acknowledged. First, this is a single-center study with internal hold-out validation only; performance estimates may therefore be optimistic, and generalizability remains to be confirmed. Given the small internal test set (n = 60), the CIs are necessarily wide. Future work will include temporal validation within our center and external, multicenter validation, with attention to site-level distribution shifts and the potential need for recalibration and threshold re-specification; larger validation cohorts should also allow narrower CIs. Second, train-test comparability was assessed using SMDs; several laboratory variables (WBC, ANC, PLT, and CRP) showed modest imbalance (|SMD| ≈ 0.21–0.39). Although none of these variables entered the final four-predictor model (high-grade malignancy, drinking, eGFR, smoking), we also conducted a prespecified sensitivity analysis in which all variables with |SMD| > 0.20 were excluded prior to preprocessing, and we reran the same pipeline. Performance was essentially unchanged relative to the primary analysis (Supplementary Tables S9, S10), suggesting limited impact of this imbalance. Nevertheless, residual bias in internal metrics cannot be fully excluded. Third, phenotype ascertainment may be constrained by the lack of systematic microbiological confirmation and by the absence of centralized, standardized radiologic re-reads/scoring. Although case adjudication was performed by clinicians based on clinical, laboratory, and imaging findings, incorporating detailed information such as pathogen identification or standardized radiographic scoring could enhance diagnostic accuracy and model performance. Fourth, despite performing feature selection and hyperparameter tuning strictly within training cross-validation folds, some risk of overfitting remains. Given the initial 35 candidates and 79 events, selection-induced instability is possible; EPV constraints, the plateau rule, nested-CV stability analyses, and LASSO bootstrapping mitigate—but do not eliminate—this risk. External (temporal/multicenter) validation will be needed to assess feature robustness and the need for recalibration, and threshold re-specification. Finally, the web tool is a prototype for internal validation and has not undergone prospective clinical implementation or impact evaluation.
Taken together, these limitations—particularly the lack of systematic microbiological confirmation, the absence of standardized radiologic re-reads/scoring, and the absence of external (temporal/multicenter) and prospective/real-world validation—substantially limit the immediate clinical translation of our findings. The model should be regarded as research-only pending external validation, potential recalibration and threshold re-specification under dataset shift, and prospective impact evaluation.
4.5 Future perspectives
Future work will prioritize temporal and external validation across broader populations and care settings. Multicenter collaborations will be essential to evaluate generalizability and transportability, mitigate overfitting, and monitor calibration drift with protocolized threshold recalibration in new cohorts. Integration with electronic health records—for research use only—will be explored to enable real-time risk scoring, subject to governance, privacy safeguards, and workflow co-design. As healthcare ML matures, future iterations may incorporate longitudinal clinical trajectories, pathogen-specific data, treatment exposures, and imaging-derived features to improve performance while maintaining interpretability.
5 Conclusion
We developed and internally validated a gradient-boosting model to estimate the 90-day risk of pneumonia after chemotherapy in NHL using routinely available predictors, with SHAP-based explanations and a research-prototype web tool. Performance on repeated cross-validation and an internal hold-out test set was encouraging; however, the single-center, retrospective design and the modest test sample warrant cautious interpretation. The tool is not intended for clinical decision-making, and external (including temporal and multicenter) validation is required before clinical deployment. With appropriate validation and updates to calibration and threshold specification, this approach may enable earlier identification of higher-risk patients and support targeted preventive strategies. All code is openly available (GitHub: https://github.com/zzn-project/NHL-Pneumonia-ML) to facilitate transparency and reproducibility.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
The studies involving humans were approved by Dongyang Hospital Affiliated with Wenzhou Medical University. The studies were conducted in accordance with the local legislation and institutional requirements. The ethics committee/institutional review board waived the requirement of written informed consent for participation from the participants or the participants’ legal guardians/next of kin because This study was conducted in accordance with the ethical principles outlined in the Declaration of Helsinki and was approved by the Ethics Committee of Dongyang Hospital Affiliated with Wenzhou Medical University (approval no. 2025-YX-132). Given the retrospective design and use of de-identified patient data, the requirement for written informed consent was waived by the ethics committee. No animal studies are presented in this manuscript, and no potentially identifiable images or data are included.
Author contributions
ZZ: Conceptualization, Data curation, Formal analysis, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing. MS: Data curation, Investigation, Writing – original draft. PJ: Data curation, Investigation, Writing – review & editing. XW: Visualization, Writing – original draft. LK: Data curation, Investigation, Writing – review & editing. XT: Writing – review & editing. GW: Conceptualization, Supervision, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by grants from the Medical and Health Science and Technology Project of Zhejiang Province (2023KY385), the Jinhua Science and Technology Bureau (2024-4-237), and the Jinhua Municipal Health Commission (2024LC03).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Gen AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fmed.2025.1674896/full#supplementary-material
References
1. Sung, H, Ferlay, J, Siegel, RL, Laversanne, M, Soerjomataram, I, Jemal, A, et al. Global Cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA A Cancer J Clinicians. (2021) 71:209–49. doi: 10.3322/caac.21660
2. Siegel, R, Naishadham, D, and Jemal, A. Cancer statistics, 2013. CA A Cancer J Clinicians. (2013) 63:11–30. doi: 10.3322/caac.21166
3. Coiffier, B, Lepage, E, Brière, J, Herbrecht, R, Tilly, H, Bouabdallah, R, et al. CHOP chemotherapy plus rituximab compared with CHOP alone in elderly patients with diffuse large-B-cell lymphoma. N Engl J Med. (2002) 346:235–42. doi: 10.1056/NEJMoa011795
4. Zhang, L, Zhang, J, He, H, Ling, X, Li, F, Yang, Z, et al. Increased cytokine levels assist in the diagnosis of respiratory bacterial infections or concurrent Bacteremia in patients with non-Hodgkin’s lymphoma. Front Cell Infect Microbiol. (2022) 12:860526. doi: 10.3389/fcimb.2022.860526
5. Periselneris, J, and Brown, JS. A clinical approach to respiratory disease in patients with hematological malignancy, with a focus on respiratory infection. Med Mycol. (2019) 57:S318–27. doi: 10.1093/mmy/myy138
6. Keefer, K, Bender, R, Liao, J, Sivik, J, and Van De Louw, A. Characteristics of pulmonary complications in non-Hodgkin’s lymphoma patients treated with rituximab-containing chemotherapy and impact on survival. Ann Hematol. (2018) 97:2373–80. doi: 10.1007/s00277-018-3448-9
7. Maschmeyer, G, De Greef, J, Mellinghoff, SC, Nosari, A, Thiebaut-Bertrand, A, Bergeron, A, et al. Infections associated with immunotherapeutic and molecular targeted agents in hematology and oncology. A position paper by the European conference on infections in Leukemia (ECIL). Leukemia. (2019) 33:844–62. doi: 10.1038/s41375-019-0388-x
8. Yang, J, Chai, L, Jia, J, Su, L, and Hao, Z. Meta-analysis of risk factors and incidence of interstitial pneumonia with CHOP-like regimens for non-Hodgkin lymphoma. Front Oncol. (2022) 12:880144. doi: 10.3389/fonc.2022.880144
9. Klastersky, J, De Naurois, J, Rolston, K, Rapoport, B, Maschmeyer, G, Aapro, M, et al. Management of febrile neutropaenia: ESMO clinical practice guidelines. Ann Oncol. (2016) 27:v111–8. doi: 10.1093/annonc/mdw325
10. Klastersky, J, Paesmans, M, Rubenstein, EB, Boyer, M, Elting, L, Feld, R, et al. The multinational association for supportive care in cancer risk index: a multinational scoring system for identifying low-risk febrile neutropenic cancer patients. J Clin Oncol. (2000) 18:3038–51. doi: 10.1200/JCO.2000.18.16.3038
11. Talcott, JA, Siegel, RD, Finberg, R, and Goldman, L. Risk assessment in cancer patients with fever and neutropenia: a prospective, two-center validation of a prediction rule. J Clin Oncol. (1992) 10:316–22. doi: 10.1200/JCO.1992.10.2.316
12. Carmona-Bayonas, A, Jiménez-Fonseca, P, Virizuela Echaburu, J, Antonio, M, Font, C, Biosca, M, et al. Prediction of serious complications in patients with seemingly stable febrile neutropenia: validation of the clinical index of stable febrile neutropenia in a prospective cohort of patients from the FINITE study. J Clin Oncol. (2015) 33:465–71. doi: 10.1200/JCO.2014.57.2347
13. Hwang, S, Kwon, KT, Kim, Y, Bae, S, Chang, H-H, Kim, S-W, et al. Usefulness analysis of the 2018 ASCO/IDSA guideline for outpatient management of fever and neutropenia in adults treated for malignancy. Sci Rep. (2021) 11:9048. doi: 10.1038/s41598-021-88207-6
14. Stanzani, M, Vianelli, N, Cavo, M, Kontoyiannis, DP, and Lewis, RE. Development and internal validation of a model for predicting 60-day risk of invasive mould disease in patients with haematological malignancies. J Infect. (2019) 78:484–90. doi: 10.1016/j.jinf.2019.04.002
15. Stanzani, M, Lewis, RE, Fiacchini, M, Ricci, P, Tumietto, F, Viale, P, et al. A risk prediction score for invasive Mold disease in patients with Hematological malignancies. PLoS One. (2013) 8:e75531. doi: 10.1371/journal.pone.0075531
16. Elhadi, M, Khaled, A, and Msherghi, A. Infectious diseases as a cause of death among cancer patients: a trend analysis and population-based study of outcome in the United States based on the surveillance, epidemiology, and end results database. Infect Agents Cancer. (2021) 16:72. doi: 10.1186/s13027-021-00413-z
17. Torres, VB, Azevedo, LC, Silva, UV, Caruso, P, Torelly, AP, Silva, E, et al. Sepsis-associated outcomes in critically ill patients with malignancies. Ann Am Thorac Soc. (2015) 12:150618124156002. doi: 10.1513/AnnalsATS.201501-046OC
18. Taplitz, RA, Kennedy, EB, Bow, EJ, Crews, J, Gleason, C, Hawley, DK, et al. Outpatient management of fever and neutropenia in adults treated for malignancy: American Society of Clinical Oncology and Infectious Diseases Society of America clinical practice guideline update. J Clin Oncol. (2018) 36:1443–53. doi: 10.1200/JCO.2017.77.6211
19. Zhu, J, Zhao, Z, Yin, B, Wu, C, Yin, C, Chen, R, et al. An integrated approach of feature selection and machine learning for early detection of breast cancer. Sci Rep. (2025) 15:13015. doi: 10.1038/s41598-025-97685-x
20. Ahn, I, Gwon, H, Kang, H, Kim, Y, Seo, H, Choi, H, et al. Machine learning–based hospital discharge prediction for patients with cardiovascular diseases: development and usability study. JMIR Med Inform. (2021) 9:e32662. doi: 10.2196/32662
21. Joel, LO, Doorsamy, W, and Paul, BS. A comparative study of imputation techniques for missing values in healthcare diagnostic datasets. Int J Data Sci Anal. (2025). doi: 10.1007/s41060-025-00825-9 [ahead of print].
22. Di Lena, P, Sala, C, Prodi, A, and Nardini, C. Missing value estimation methods for DNA methylation data. Bioinformatics. (2019) 35:3786–93. doi: 10.1093/bioinformatics/btz134
24. Austin, PC. Balance diagnostics for comparing the distribution of baseline covariates between treatment groups in propensity-score matched samples. Stat Med. (2009) 28:3083–107. doi: 10.1002/sim.3697
25. Stuart, EA. Matching methods for causal inference: a review and a look forward. Stat Sci. (2010) 25:1–21. doi: 10.1214/09-STS313
26. Liu, J, Ma, Y, Xie, W, Li, X, Wang, Y, Xu, Z, et al. Lasso-based machine learning algorithm for predicting postoperative lung complications in elderly: a single-center retrospective study from China. Clin Interv Aging. (2023) 18:597–606. doi: 10.2147/CIA.S406735
27. Yang, T, Huang, L, He, J, Luo, L, Guo, W, Chen, H, et al. Establishment of diagnostic model and identification of diagnostic markers between liver cancer and cirrhosis based on multi-chip and machine learning. Clin Exp Pharma Physio. (2024) 51:e13907. doi: 10.1111/1440-1681.13907
28. Lee, YW, Choi, JW, and Shin, E-H. Machine learning model for predicting malaria using clinical information. Comput Biol Med. (2021) 129:104151. doi: 10.1016/j.compbiomed.2020.104151
29. Sun, J, Sun, C-K, Tang, Y-X, Liu, T-C, and Lu, C-J. Application of SHAP for explainable machine learning on age-based subgrouping mammography questionnaire data for positive mammography prediction and risk factor identification. Healthcare. (2023) 11:2000. doi: 10.3390/healthcare11142000
30. Wang, Y, Lei, J, Jin, Z, Jiang, Y, Zhang, N, Lv, M, et al. Development and validation of a machine learning-based clinical prediction model for monitoring liver injury in patients with pan-cancer receiving immunotherapy. Int J Med Inform. (2025) 203:106036. doi: 10.1016/j.ijmedinf.2025.106036
31. Jiang, W, Chen, H, Yang, L, and Pan, X. moreThanANOVA: a user-friendly shiny/R application for exploring and comparing data with interactive visualization. PLoS One. (2022) 17:e0271185. doi: 10.1371/journal.pone.0271185
32. Teng, X, Han, K, Jin, W, Ma, L, Wei, L, Min, D, et al. Development and validation of an early diagnosis model for bone metastasis in non-small cell lung cancer based on serological characteristics of the bone metastasis mechanism. eClinicalMedicine. (2024) 72:102617. doi: 10.1016/j.eclinm.2024.102617
33. Metlay, JP, Waterer, GW, Long, AC, Anzueto, A, Brozek, J, Crothers, K, et al. Diagnosis and treatment of adults with community-acquired pneumonia. An official clinical practice guideline of the American Thoracic Society and Infectious Diseases Society of America. Am J Respir Crit Care Med. (2019) 200:e45–67. doi: 10.1164/rccm.201908-1581ST
34. Sun, T, Liu, J, Yuan, H, Li, X, and Yan, H. Construction of a risk prediction model for lung infection after chemotherapy in lung cancer patients based on the machine learning algorithm. Front Oncol. (2024) 14:1403392. doi: 10.3389/fonc.2024.1403392
35. Peng, T, Liu, L, Liu, F, Ding, L, Liu, J, Zhou, H, et al. Machine learning-based infection prediction model for newly diagnosed multiple myeloma patients. Front Neuroinform. (2023) 16:1063610. doi: 10.3389/fninf.2022.1063610
36. Li, C, Lu, F, Lei, T, Yu, H, Chen, X, Peng, S, et al. Prophylactic antibiotic treatment with TMP-SMX decreased the incidence of interstitial pneumonia in patients with B-cell lymphoma on chemotherapy. BMC Cancer. (2020) 20:742. doi: 10.1186/s12885-020-07254-w
37. Zhao, H, Su, R, Luo, J, and Liu, L. Incidence and risk factors of severe infections in diffuse large B-cell lymphoma patients undergoing immunochemotherapy. Ann Hematol. (2025) 104:2869–79. doi: 10.1007/s00277-025-06296-z
38. Dendle, C, Gilbertson, M, Spelman, T, Stuart, RL, Korman, TM, Thursky, K, et al. Infection is an independent predictor of death in diffuse large B cell lymphoma. Sci Rep. (2017) 7:4395. doi: 10.1038/s41598-017-04495-x
39. Zhao, J, Zhang, Y, Wang, W, Zhang, W, and Zhou, D. Post-chemotherapy pneumonia in Chinese patients with diffuse large B-cell lymphoma: outcomes and predictive model. Front Oncol. (2022) 12:955535. doi: 10.3389/fonc.2022.955535
40. Wang, J, Liu, F, and Tang, X. Incidence and risk factors of pneumonia in diffuse large B-cell lymphoma patients receiving first line R-CHOP/R-CHOP-like immunochemotherapy: a retrospective study of 287 patients in single center. Ann Palliat Med. (2021) 10:11931–8. doi: 10.21037/apm-21-3280
41. Simou, E, Britton, J, and Leonardi-Bee, J. Alcohol and the risk of pneumonia: a systematic review and meta-analysis. BMJ Open. (2018) 8:e022344. doi: 10.1136/bmjopen-2018-022344
42. Taj, M, Nadeem, M, Maqsood, S, Shah, T, Farzana, T, and Shamsi, TS. Validation of MASCC score for risk stratification in patients of Hematological disorders with febrile neutropenia. Indian J Hematol Blood Transfus. (2017) 33:355–60. doi: 10.1007/s12288-016-0730-7
43. Sheehy, J, Gallanagh, M, Sullivan, C, and Lane, S. Clinical prediction models for febrile neutropenia and its outcomes: a systematic review. Support Care Cancer. (2025) 33:537. doi: 10.1007/s00520-025-09562-y
44. Xu, L, Li, C, Zhang, J, Guan, C, Zhao, L, Shen, X, et al. Personalized prediction of mortality in patients with acute ischemic stroke using explainable artificial intelligence. Eur J Med Res. (2024) 29:341. doi: 10.1186/s40001-024-01940-2
Glossary
NHL - Non-Hodgkin Lymphoma
RF-RFE - Random Forest-based Recursive Feature Elimination
DLBCL - Diffuse Large B-cell Lymphoma
ML - Machine Learning
LR - Logistic Regression
SVM - Support Vector Machine
KNN - k-Nearest Neighbors
GBM - Gradient Boosting Machine
LightGBM - Light Gradient Boosting Machine
AUC - Area Under the Curve
DCA - Decision Curve Analysis
SHAP - Shapley Additive Explanations
NCCN-IPI - National Comprehensive Cancer Network-International Prognostic Index
WBC - White blood cell count
Hb - Hemoglobin
ANC - Absolute neutrophil count
ALC - Absolute lymphocyte count
RDW - Red cell distribution width
PLT - Platelet count
MPV - Mean platelet volume
PCT - Plateletcrit
eGFR - Estimated glomerular filtration rate
Ca - Calcium
Fe - Iron
ALB - Albumin
A/G ratio - Albumin/globulin ratio
ALT - Alanine aminotransferase
AST - Aspartate aminotransferase
LDH - Lactate dehydrogenase
α-HBDH - α-Hydroxybutyrate dehydrogenase
TG - Triglycerides
LDL - Low-density lipoprotein cholesterol
CRP - C-reactive protein
Keywords: non-Hodgkin lymphoma, pneumonia, machine learning, risk prediction, Shapley additive explanations
Citation: Zhang Z, Su M, Jiang P, Wang X, Kong L, Tong X and Wu G (2025) Interpretable machine learning model to predict 90-day radiographically confirmed pneumonia after chemotherapy initiation in non-Hodgkin lymphoma: development and internal validation of a single-center cohort. Front. Med. 12:1674896. doi: 10.3389/fmed.2025.1674896
Edited by:
Wenlin Yang, University of Florida, United StatesReviewed by:
Victor Hugo Nieto Estrada, El Bosque University, ColombiaHuasheng Lv, Xinjiang Medical University, China
Copyright © 2025 Zhang, Su, Jiang, Wang, Kong, Tong and Wu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Xiangmin Tong, dG9uZ3hpYW5nbWluQDE2My5jb20=; Gongqiang Wu, d3Vnb25ncWlhbmc1OUAxMjYuY29t
†These authors have contributed equally to this work