Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Comput. Neurosci., 18 August 2025

Volume 19 - 2025 | https://doi.org/10.3389/fncom.2025.1615576

Transformer-based multimodal precision intervention model for enhancing diaphragm function in elderly patients


Ma XinliMa Xinli1Zhao JieZhao Jie1Yan MingYan Ming2Zhang YanpingZhang Yanping1Li FanLi Fan3Jia JingJia Jing3Ding Lu
Ding Lu4*
  • 1Critical Medicine Department, The Second Hospital of Jilin University, Changchun, Jilin, China
  • 2Nursing Department, The Second Hospital of Jilin University, Changchun, Jilin, China
  • 3School of Nursing, Jilin University, Changchun, Jilin, China
  • 4General Surgery Department, The Second Hospital of Jilin University, Changchun, Jilin, China

Diaphragm dysfunction represents a significant complication in elderly patients undergoing mechanical ventilation, often resulting in extended intensive care stays, unsuccessful weaning attempts, and increased healthcare expenditures. To address the deficiency of precise, real-time decision support in this context, a novel artificial intelligence framework is proposed, integrating imaging, physiological signals, and ventilator parameters. Initially, a hierarchical Transformer encoder is employed to extract modality-specific embeddings, followed by an attention-guided cross-modal fusion module and a temporal network for dynamic trend prediction. The framework was assessed using three public datasets, which are, the MIMIC-IV, eICU, and Chest X-ray. The proposed model achieved the highest accuracy (92.3% on MIMIC-IV, 91.8% on eICU, 92.0% on Chest X-ray) and surpassed all baselines in precision, recall, F1-score, and Matthews correlation coefficient. Additionally, the model's probability estimates were well-calibrated, and its SHAP-based explainability analysis identified ventilator volume and key imaging features as primary predictors. The clinical implications of this study are significant. By providing precise and interpretable predictions, the proposed model has the potential to transform critical care practices by offering a pathway to more effective and personalized interventions for high-risk patients.

1 Introduction

Diaphragm dysfunction constitutes a significant clinical challenge in the management of elderly patients, particularly those who require mechanical ventilation (Kaya and Iscimen, 2025). As the primary muscle responsible for respiration, the diaphragm's optimal function is crucial for effective breathing and gas exchange (Lessa et al., 2016). However, in elderly populations, age-related muscle degeneration, comorbidities, and the detrimental effects of prolonged mechanical ventilation contribute to a high incidence of diaphragm dysfunction (Powers et al., 2009; Siniscalchi et al., 2024). This dysfunction not only compromises respiratory efficiency but also elevates the risk of extended intensive care unit (ICU) stays, ventilator-associated complications, and overall mortality (Ganatra and Varisco, 2019).

Mechanical ventilation, while lifesaving, presents several inherent challenges that complicate the assessment of diaphragm function (Ratano et al., 2024). Ventilator-induced diaphragm dysfunction (VIDD) is a well-documented phenomenon, wherein the act of mechanical support leads to diaphragmatic atrophy and impaired contractility (Vassilakopoulos and Petrof, 2004). Traditional methods of diaphragm evaluation, such as ultrasound imaging and electromyography, are often operator-dependent and limited in their capacity to capture the complex, dynamic interplay between patient physiology and ventilatory support (Laghi et al., 2021; Sarwar et al., 2024). These limitations hinder timely intervention and may result in suboptimal management strategies in critically ill patients.

Current clinical practices and predictive models for assessing diaphragm function exhibit significant limitations (Llamas-Álvarez et al., 2017). Most conventional models rely heavily on static snapshots of clinical parameters, neglecting the temporal evolution of respiratory mechanics (Hsia et al., 2023; Hantos, 2021). Additionally, many existing methods fail to integrate the heterogeneous nature of available clinical data, ranging from imaging and physiological parameters to ventilator settings, thereby restricting their predictive accuracy (Howroyd et al., 2024). The reliance on isolated, single-modality data often results in models that are unable to capture the nuanced and dynamic patterns inherent in patient responses during mechanical ventilation (Adibi et al., 2025). Consequently, clinicians are left with decision-support tools that provide limited insights and lack the robustness necessary for effective intervention planning.

The primary research question of this study is: Can a Transformer-based multimodal model that simultaneously processes imaging, physiological signals, and ventilator metrics accurately predict diaphragm dysfunction in elderly patients undergoing mechanical ventilation and facilitate personalized intervention planning? Surpassing current methodologies, which typically concentrate on individual data streams or static snapshots, the proposed framework distinctively integrates hierarchical Transformer encoding, cross-modal attention fusion, and temporal modeling. Additionally, it offers interpretable explanations to provide real-time, data-driven decision support.

Despite advancements in machine learning for critical care (Zhang et al., 2024; Biesheuvel et al., 2024), a significant research gap remains: no current methodology concurrently addresses the temporal dynamics of diaphragm function, integrates multimodal clinical data at scale, and ensures interpretability for clinicians. Previous studies have either concentrated on individual data streams (e.g., imaging alone) or utilized traditional time-series models that do not capitalize on recent deep learning architectures. Notably, Transformer-based methods, which are proficient in modeling long-range dependencies in heterogeneous data, are underutilized in this field (Kim et al., 2025). This gap highlights the necessity for a comprehensive, real-time predictive framework capable of integrating imaging, physiological signals, and ventilator metrics to inform personalized interventions.

The emergence of artificial intelligence (AI) and deep learning presents a transformative potential to address these challenges. In recent years, deep learning architectures, particularly those utilizing Transformer models, have demonstrated significant success in processing complex and multimodal data (Xu et al., 2023; Zhang et al., 2023; Mostafaei et al., 2024). Transformer models, characterized by their multi-head self-attention mechanisms, excel at capturing both local and global dependencies within data sequences, rendering them highly suitable for time-series analysis in critical care environments (Ma et al., 2019b; Yamazaki, 2024). The capability of these models to integrate information from diverse sources, such as imaging data, clinical text, and physiological signals, offers a unique opportunity to develop more comprehensive and accurate predictive models for diaphragm dysfunction.

This study presents a novel Transformer-based multimodal framework designed to integrate diverse clinical data, thereby addressing the previously identified limitations. The framework employs a hierarchical Transformer encoder for feature extraction, which is succeeded by an attention-guided fusion module that aligns imaging and textual features. Subsequently, a temporal network captures dynamic patient trends, facilitating predictions that support early and personalized interventions.

The contributions of this work are 3fold.

• First, an AI-driven framework is developed to predict diaphragm dysfunction in elderly patients under mechanical ventilation by integrating diverse data sources.

• Second, the clinical utility of the framework is validated against state-of-the-art methods using rigorous performance and responsiveness metrics.

• Third, the framework's capacity for personalized intervention planning is demonstrated, showing how dynamic ventilator adjustments mitigate diaphragm atrophy and improve weaning outcomes.

The rest of this work is organized as follows. In Section 2, a comprehensive review of the literature on diaphragm dysfunction and mechanical ventilation and transformer models is presented. Section 3 describes the data processing strategy used in this work. The methdological description of proposed work is presented in Section 4. In Section 5, the model training and validation criteria are described. The evaluation porotocols are presented in Section 6. The comprehensive presentation of the experimental results is given in Section 7. Finally, the conclusions are stated in Section 8.

2 Literature review

2.1 Diaphragm dysfunction and mechanical ventilation

Diaphragm dysfunction in patients undergoing mechanical ventilation arises from a multitude of interconnected mechanisms. Prolonged mechanical ventilation can result in VIDD, which is characterized by diaphragmatic atrophy and impaired contractility (Panelli et al., 2024). This condition is attributed to reduced diaphragmatic activity, oxidative stress, and the activation of proteolytic pathways, ultimately leading to muscle fiber degradation and a decreased capacity for force generation. Scott K Power (Powers, 2024) emphasized that controlled mechanical ventilation can rapidly induce diaphragmatic atrophy and weakness due to inactivity and oxidative stress. Furthermore, sepsis and systemic inflammation exacerbate diaphragmatic weakness by promoting protein degradation and mitochondrial dysfunction.

Diaphragm dysfunction exerts a significant impact on patient outcomes, particularly in relation to weaning from mechanical ventilation (Battaglini and Rocco, 2024). Patients with diaphragmatic weakness frequently experience prolonged weaning periods, increased rates of extubation failure, and elevated mortality (Yayan and Schiffner, 2024). Research has demonstrated that diaphragm weakness is associated with extended durations of mechanical ventilation and prolonged ICU stays, thereby increasing healthcare costs and resource utilization. For example, a study by Jung et al. (2016) found that diaphragm dysfunction in ICU patients was linked to higher rates of extubation failure and prolonged mechanical ventilation.

2.2 AI in critical care

AI has been increasingly integrated into critical care to enhance patient monitoring, predict clinical deterioration, and support decision-making (Yoon et al., 2022; Bhakar et al., 2023). AI applications encompass predictive analytics for sepsis detection, mortality risk assessment, and optimization of treatment protocols (Yang et al., 2023). Machine learning models have been employed to analyze complex datasets, facilitating early detection of patient deterioration and informing therapeutic interventions.

In the realm of respiratory care, AI models have been developed to predict respiratory failure, optimize ventilator settings, and assess readiness for weaning (Ahmed et al., 2025). These models utilize patient data, including vital signs, laboratory results, and ventilator parameters, to forecast respiratory decline and guide clinical decisions (Liao et al., 2021). For example, AI algorithms have been applied to predict extubation outcomes by analyzing patterns in respiratory parameters and patient characteristics (Torrini et al., 2021).

Recent investigations have underscored the efficacy of hybrid and metaheuristic-tuned AI models in the realm of medical diagnostics. A notable example is a hybrid convolutional neural network (CNN) integrated with XGBoost, whose hyperparameters were optimized using a modified Arithmetic Optimization Algorithm, achieving an impressive accuracy of 99.39% on a balanced COVID-19 chest X-ray dataset comprising 12,000 images (Zivkovic et al., 2022). In the domain of audio-based respiratory condition classification, the use of CNNs, XGBoost, and AdaBoost, fine-tuned by a modified Particle Swarm Optimization, yielded an accuracy of 98.14% in binary detection and 81.25% in multiclass scenarios (Purkovic et al., 2024). Furthermore, a metaheuristic-optimized Long Short-Term Memory network applied to gait time series data demonstrated an accuracy of 89.92% in detecting Parkinson's disease (Markovic et al., 2025). Similarly, a CNN designed for respiratory condition detection from audio, optimized by a bespoke metaheuristic algorithm, achieved 93% accuracy in binary classification and 75% in multiclass identification (Bacanin et al., 2024a). Additionally, anomaly detection in ECG signals using Recurrent Neural Networks, optimized with the Crayfish Optimization Algorithm, demonstrated superior performance in the early detection of cardiovascular disorders (Jovanovic et al., 2023).

Another recent research has investigated the integration of metaheuristic optimization with deep learning to enhance diagnostic accuracy in complex biomedical applications. For example, Bacanin et al. (2024b) developed a model for detecting Parkinson's disease by combining a deep Long Short-Term Memory (LSTM) network with a modified metaheuristic algorithm for hyperparameter tuning. This study utilizes gait time-series data, employing the metaheuristic to optimize LSTM parameters such as learning rate, hidden layer size, and sequence length. The optimized network achieved an accuracy of 89.92%, indicating that metaheuristic-guided tuning can significantly improve model performance in the detection of neurodegenerative diseases. Despite advancements, existing AI approaches in critical care face several limitations. Many models operate as “black boxes,” lacking transparency in their decision-making processes, which hinders clinician trust and adoption (Marey et al., 2024). Additionally, the integration of heterogeneous data sources remains challenging, often resulting in models that do not fully capture the complexity of patient physiology (Xu et al., 2024). Moreover, the generalizability of AI models is frequently limited by dataset biases and variability across different clinical settings.

2.3 Transformer models and multimodal data fusion

Originally developed for natural language processing (Vaswani et al., 2017), transformer models have exhibited significant advantages in handling sequential data, primarily due to their self-attention mechanisms (Ma et al., 2019a; Kumar and Solanki, 2023). These models excel in capturing long-range dependencies and contextual relationships within data sequences, making them particularly well-suited for analyzing time-series clinical data (Jiang et al., 2024). The ability to process entire sequences concurrently allows Transformers to effectively model complex temporal patterns in patient data (Ganiya et al., 2024).

The integration of multimodal data, including clinical notes, imaging studies, and physiological signals, is crucial for a comprehensive patient assessment (Xu et al., 2024). Traditional data fusion methods frequently encounter challenges in aligning and interpreting diverse data types (Duan et al., 2024). Transformer-based models offer a promising solution by providing a unified framework for multimodal integration, employing self-attention mechanisms to dynamically evaluate the significance of each data modality (Rashno et al., 2024). This approach enhances the model's ability to capture the nuanced interactions between different data sources, leading to more accurate predictions and insights.

While Transformer models have demonstrated potential across various domains, their application within the healthcare sector, particularly for the fusion of multimodal data, remains underexplored (AlSaad et al., 2024). Challenges such as managing missing data, ensuring interpretability, and maintaining computational efficiency must be addressed. Future research should concentrate on developing Transformer-based architectures specifically designed to accommodate the complexities of clinical data, integrating domain knowledge to enhance model performance and reliability. Additionally, strategies to improve model transparency and clinician interpretability are essential for successful integration into clinical practice.

3 Data processing

3.1 Data collection and preprocessing

This study utilizes a diverse range of multimodal datasets to facilitate a comprehensive analysis of diaphragm function in elderly patients undergoing mechanical ventilation. The data sources encompass clinical records, physiological parameters, imaging data, and ventilator data. Clinical records provide information on patient demographics, medical histories, and laboratory test results, while physiological parameters capture real-time vital signs such as heart rate, blood pressure, and oxygen saturation. Additionally, imaging data, including high-resolution chest X-rays, offer visual insights into pulmonary status and diaphragm positioning, and ventilator data include detailed settings and performance metrics essential for assessing ventilation dynamics.

Ethical considerations were integral to the study design. All patient data were anonymized in accordance with the guidelines established by the Institutional Review Board (IRB) and in compliance with the Health Insurance Portability and Accountability Act (HIPAA) regulations. Each dataset underwent rigorous de-identification to ensure the removal of personal identifiers prior to data analysis. Where applicable, informed consent was obtained from patients or their legal representatives, ensuring adherence to ethical standards for research involving human subjects.

Data preprocessing involved several key steps aimed at enhancing the quality and integrity of the raw data. Initially, the datasets underwent a thorough data cleaning process, which included the removal of duplicates, correction of errors, and imputation of missing values. In instances where missing data were minimal, mean or median imputation was applied; however, for larger gaps, advanced imputation techniques such as multiple imputation by chained equations (MICE) were employed. Outlier detection was conducted using both statistical methods and domain knowledge, ensuring that extreme values, which could potentially skew the analysis, were appropriately addressed.

Normalization and standardization of data are imperative for the integration of multimodal inputs. For continuous variables, z-score normalization was employed, mathematically represented as:

x=x-μσ,    (1)

where x denotes an individual measurement, μ is the mean of the dataset, and σ is the standard deviation. This procedure is crucial to ensure that the varying scales of clinical, imaging, and ventilator data do not bias the feature extraction process of the AI model. One-hot encoding was applied for categorical data, converting categorical variables into a format readily processed by deep learning algorithms. An additional preprocessing step involved aligning the temporal dimensions of the data. Given that ventilator settings and physiological measurements are time-series data, these records were synchronized with the timestamps of imaging studies to ensure that the multimodal inputs accurately corresponded to the same clinical events. This temporal alignment is particularly important for models utilizing Transformer architectures, as it enhances the model's ability to capture dynamic changes over time.

3.2 Public datasets integration

To ensure the robustness and generalizability of the model, data from three widely recognized public databases were integrated. The first is the MIMIC-IV Clinical Database, which offers a comprehensive collection of critical care clinical records and physiological data (Johnson et al., 2023). The second, the eICU Collaborative Research Database, provides a multi-center repository of ICU records with detailed patient monitoring data (Pollard et al., 2018). The third dataset consists of chest X-ray images specifically curated for pneumonia detection (Kermany, 2018). Table 1 summarizes the key attributes of these datasets.

Table 1
www.frontiersin.org

Table 1. Summary of integrated public datasets.

4 Proposed AI-driven framework

This section introduces the innovative AI-driven framework designed to integrate heterogeneous clinical data modalities for predicting and monitoring diaphragm function in elderly patients undergoing mechanical ventilation. The framework comprises a hierarchical Transformer-based feature extraction module, an attention-guided cross-modal fusion module, and a temporal network module. Each component is meticulously crafted to address the specific challenges associated with multimodal clinical data integration and to capture both static and dynamic aspects of patient physiology. Figure 1 presents proposed framework which integrates the multimodal clinical data through a hierarchical Transformer encoder, attention-guided cross-modal fusion, and a temporal network module, culminating in real-time prediction and intervention outputs.

Figure 1
Diagram of a machine learning model processing ventilator, physiological, and imaging data using convolutional layers and low feature extraction (LFE). It includes a hierarchical transformer encoder (HTE) and decoder, followed by transformer features. The model outputs prediction with an attention-guided cross-modal fusion and temporal network module. The prediction output shows a graph of predicted dysrhythmia probability with uncertainty zone intervals. The components are labeled and color-coded for clarity.

Figure 1. Overall architecture of the proposed AI-driven framework for enhancing diaphragm function in elderly patients.

In Figure 1, each input modality, including imaging, physiological, and ventilator data, is transformed into a 512-dimensional embedding. The images undergo processing via a pretrained ResNet-50, are subsequently flattened, and then linearly mapped to ℝ512. Time-series features are concatenated and processed through a linear layer to achieve the same dimensionality. Positional encodings are incorporated, and the sequence of length T is input into N = 6 Transformer layers. Each layer employs 8-head self-attention with a key/query/value dimension of 64 (dmodel/8) and a two-layer feed-forward network comprising 2048 hidden units with ReLU activation. Dropout (p = 0.1), residual connections, and layer normalization are utilized to ensure stable training and robust feature extraction.

4.1 Transformer-based feature extraction

At the core of this framework lies a hierarchical Transformer encoder, which is proficient in learning contextual representations from heterogeneous data. This design draws inspiration from the success of Transformer architectures in natural language processing, yet it is specifically adapted for clinical data. The encoder consists of multiple layers of multi-head self-attention (MHSA) mechanisms and position-wise feed-forward networks. Each layer processes the input features to progressively capture fine-grained local interactions at the lower levels and abstract global dependencies at higher levels. The following equation defines the self-attention mechanism:

Attention(Q,K,V)=softmax(QKTdk)V,    (2)

where Q (query), K (key), and V (value) matrices are linear projections of the input features, and dk represents the dimensionality of the key vectors. This formulation enables the network to dynamically weigh the importance of different features, facilitating the extraction of nuanced patterns across diverse clinical and imaging datasets. To further enhance the feature representation, the multi-head self-attention mechanism is employed, mathematically expressed as:

MultiHead(Q,K,V)=Concat(head1,,headh)WO,    (3)

with each head computed as:

headi=Attention(QWiQ,KWiK,VWiV),    (4)

where WiQ, WiK, WiV, and WO are learnable projection matrices. The use of multiple heads offers a dual advantage: it allows the model to focus on different aspects of the data simultaneously and mitigates the risk of overfitting by ensuring a diverse set of representations. This hierarchical approach provides a robust foundation for downstream tasks, particularly when fusing heterogeneous data sources. The attention-guided cross-modal fusion module is presented in Figure 2 for a better readership.

Figure 2
Diagram depicting a cross-modal attention model. The left section shows imaging and textual features processed through convolutional and fully connected layers. The centralized section combines these to produce transfer learning features. The right side presents a heat map illustrating cross-modal attention, with attention weights ranging from 0.02 to 0.16 across imaging regions and text tokens.

Figure 2. The demonstration of how imaging and textual features are aligned and fused using an attention mechanism to generate a unified representation that drives downstream predictive analytics.

4.2 Cross-modal fusion module

The integration of imaging data with clinical text features poses a significant challenge due to the intrinsic differences in their data representations. The cross-modal fusion module is specifically designed to align and integrate these diverse modalities into a unified latent space, thereby enabling a synergistic analysis that capitalizes on the strengths of each data type. Let FIn×d represent the feature matrix extracted from imaging data, and FTm×d represent the features derived from clinical text data. To align these modalities, an attention-based mechanism is employed to calculate the similarity between the two sets of features:

Ffused=softmax(FIFTTd)FT.    (5)

This operation effectively generates attention weights that highlight the most informative interactions between the imaging and text features. Additionally, a multimodal embedding layer is implemented, which is trained to minimize the semantic distance between correlated features from different modalities. This embedding process not only aligns the modalities but also facilitates the discovery of latent correlations that might otherwise remain obscured. The cross-modal fusion is further refined using an attention-based aggregation method that emphasizes salient features while suppressing noise. This is achieved by iteratively applying a weighted combination of the aligned features, thereby ensuring that both modality-specific and shared information are preserved. The resulting fused representation is then utilized as a comprehensive descriptor for the patient's condition.

4.3 Temporal network module

A critical aspect of monitoring diaphragm function is capturing the temporal evolution of patient data. To this end, the framework integrates a temporal network module that models sequential data trends over time. The outputs from the hierarchical Transformer encoder, denoted as H = {h1, h2, …, hT}, are fed into a gated recurrent unit (GRU) network, which is selected for its efficiency and ability to capture long-term dependencies. Figure 3 shows the representation of sequential processing in proposed framework.

Figure 3
Diagram of a sequence of connected nodes and boxes, representing a chain structure. Circles labeled Z-one to Z-t are positioned above rectangles labeled H-one to H-t. Circles labeled S-one to S-t are below, with arrows indicating flow between these elements in a horizontal direction.

Figure 3. The sequential processing of hidden states, capturing dynamic trends over time to produce accurate predictions and timely intervention recommendations.

The GRU network is governed by the following equations:

zt=σ(Wzht+Uzst-1+bz),    (6)
rt=σ(Wrht+Urst-1+br),    (7)
s~t=tanh(Whht+Uh(rtst-1)+bh),    (8)
st=(1-zt)st-1+zts~t,    (9)

where st is the hidden state at time t, zt is the update gate, rt is the reset gate, and ⊙ denotes element-wise multiplication. These equations enable the GRU to adaptively filter and retain essential information over time, effectively capturing temporal trends in the patient's physiological parameters and treatment responses. A concatenation strategy is employed to integrate the temporal information with the static, high-level features extracted by the Transformer encoder, followed by a fully connected layer. This layer projects the combined features into a prediction space, facilitating trend analysis and real-time decision-making for personalized ventilator adjustments. The fusion of temporal and static features enhances predictive accuracy and provides a dynamic view of patient status, which is critical for timely intervention.

4.4 Novel contributions and technical rationale

The proposed framework constitutes a significant advancement in the domain of AI within critical care by addressing several pivotal challenges. Firstly, the hierarchical Transformer encoder is distinctively tailored to process multimodal clinical data, adeptly capturing both local and global dependencies through its layered architecture and multi-head self-attention mechanisms. Secondly, the cross-modal fusion module employs attention-based techniques to align imaging and clinical text features, thereby effectively bridging the semantic gap between these disparate data types. Lastly, the temporal network module, driven by GRUs, ensures that the dynamic progression of patient data is modeled with high fidelity, facilitating precise trend predictions that are essential for adaptive clinical interventions. Mathematically, the framework exploits the synergy between attention mechanisms and recurrent neural networks. The attention operations in both the Transformer and cross-modal fusion modules enable the selective integration of pertinent features, while the GRU equations ensure that temporal dependencies are preserved without incurring excessive computational costs. This combination results in a robust, scalable model capable of managing the complex, high-dimensional data characteristic of contemporary clinical environments.

5 Model training and validation

The model's training process was specially designed to optimize performance while ensuring robust generalization across diverse clinical datasets. Table 2 provides a summary of the key hyperparameters used in the training process. Detailed performance metric analyses, including sensitivity, specificity, and other evaluation scores, will be provided in a subsequent sections.

Table 2
www.frontiersin.org

Table 2. Details of hyperparameters.

Hyperparameter optimization was systematically performed using a grid search approach. A range of values for critical parameters was explored, including dropout rates between 0.1 and 0.3, and the number of Transformer layers varying from 4 to 8. The final configuration was selected based on the optimal validation loss observed during cross-validation, ensuring that the model was tuned for both performance and computational efficiency.

To evaluate model robustness, a 5-fold cross-validation strategy was adopted. The dataset was randomly partitioned into five subsets, where in each fold, 80% of the data was used for training and 20% for validation. This approach not only enhances the reliability of the model's performance but also helps identify potential overfitting issues and ensures that the model is generalizable to unseen data.

6 Evaluation criteria

The proposed model is evaluated using a comprehensive array of metrics to ensure robust performance and reliability across all predictive aspects. The evaluation considers Accuracy, Sensitivity, Specificity, Area Under the Receiver Operating Characteristic Curve (AUC), Precision, F1 Score, Matthews Correlation Coefficient (MCC), and calibration measured via the Brier score. Each metric provides unique insights into the model's strengths and weaknesses, particularly in the context of clinical decision support.

6.1 Accuracy

The accuracy is defined as the ratio of correctly classified instances to the total number of instances:

Accuracy=TP+TNTP+TN+FP+FN,    (10)

where TP and TN denote true positives and true negatives, respectively, and FP and FN denote false positives and false negatives.

6.2 Sensitivity

This metric is also known as the true positive rate or recall, quantifies the model's ability to correctly identify positive cases:

Sensitivity=TPTP+FN.    (11)

This metric is critical in clinical settings where missing a positive case could have significant adverse outcomes.

6.3 Specificity

In contrast to sensitivity, the specificity measures the model's ability to correctly identify negative cases:

Specificity=TNTN+FP.    (12)

A high specificity indicates a low rate of false positives, thereby reducing the risk of unnecessary clinical interventions.

6.4 Precision

This metric is alternatively known as the positive predictive value which is defined as:

Precision=TPTP+FP,    (13)

which reflects the proportion of positive predictions that are correct. In practice, a high precision value is essential to ensure that positive diagnoses are accurate.

6.5 F1 score

To balance precision and sensitivity, the F1 score, the harmonic mean of precision and sensitivity, is computed:

F1 Score=2×Precision × Sensitivity Precision + Sensitivity.    (14)

The F1 Score is particularly useful in situations where the class distribution is imbalanced, as it accounts for both false positives and false negatives.

6.6 Matthews correlation coefficient (MCC)

This metric provides a balanced measure that can be used even when the classes are of very different sizes. It is defined as:

MCC=TP×TN-FP×FN(TP+FP)(TP+FN)(TN+FP)(TN+FN).    (15)

An MCC value of +1 indicates perfect prediction, 0 indicates a performance no better than random guessing, and -1 indicates total disagreement between predictions and observations.

6.7 AUC-ROC

This evaluation method is a robust indicator of the model's discriminative ability. Although AUC is typically computed using numerical integration methods (such as the trapezoidal rule) applied to the ROC curve—which plots the true positive rate against the false positive rate—it fundamentally represents the probability that a randomly selected positive instance is ranked higher than a randomly selected negative instance.

6.8 Brier score

Finally, the model's calibration is assessed with the Barier score, which quantifies the mean squared difference between predicted probabilities and the actual binary outcomes:

Brier Score=1Ni=1N(pi-yi)2,    (16)

In this context, pi represents the predicted probability for instance i, yi denotes the corresponding true label, and N signifies the total number of predictions. A lower Brier score suggests that the model's probability estimates are well-calibrated.

7 Results

7.1 Predictive performance analysis

The predictive performance of the proposed AI-driven framework was evaluated using three clinical datasets: MIMIC-IV, eICU, and Chest X-ray Images. Figure 4 illustrates the composite training loss curve for the MIMIC-IV dataset, demonstrating a rapid reduction in loss from ~1.0 in the initial epochs to below 0.15 by epoch 50. Similar trends are evident in Figures 5, 6 for the eICU and Chest X-ray datasets, respectively.

Figure 4
Line graph titled “Training Loss Curve on MIMIC-IV” showing training loss across fifty epochs. Seven models are compared, with each model represented by different colored lines. All models show a decrease in training loss, converging toward zero as epochs increase.

Figure 4. Training loss evolution on the MIMIC-IV dataset, demonstrating rapid convergence of proposed method compared to baselines.

Figure 5
Line graph titled “Training Loss Curve on eICU” showing training loss over 50 epochs for various models. Different colored lines represent models by the proposed method and several researchers from 2023 to 2025. All models show a decreasing trend in loss, converging near zero after 50 epochs.

Figure 5. Training loss progression on the eICU dataset, illustrating smooth convergence across training epochs.

Figure 6
Line graph titled “Training Loss Curve on Chest X-ray Images” showing training loss over 50 epochs for multiple models. All models, represented by different colors, show a decreasing trend, indicating improved performance over time. The legend identifies each model, including a proposed model and those by Khader et al. (2023), Kumar et al. (2024 and 2025), Xie et al. (2023), and Narmadha et al. (2025).

Figure 6. Training loss trend on the Chest X-ray dataset, where rapid loss reduction demonstrates efficient learning by the proposed framework.

In contrast, competing methods (Khader et al., 2023, Kumar and Sharma, 2024, and Kumar et al., 2025) achieved final training losses ranging from 0.20 to 0.25, indicating a slower convergence rate. Validation loss curves further confirm these findings. As depicted in Figures 79, the proposed method consistently outperforms the baselines, with final validation losses around 0.18, compared to 0.22–0.27 observed for the other methods. This clear divergence between training and validation metrics confirms that the model generalizes well, effectively minimizing overfitting despite the complexity of the multimodal data. The ROC curves presented in Figures 1012 demonstrate the high discriminative power of the framework.

Figure 7
Line graph titled “Validation Loss Curve on MIMIC-IV” showing validation loss against epochs. Six models are compared: Proposed model, Khader et al. (2023), Kumar et al. (2024, 2025), Xie et al. (2023), and Narmadha et al. (2025). All models show a decreasing trend in validation loss over 50 epochs.

Figure 7. Validation loss trends on the MIMIC-IV dataset, underscoring the robust generalization capabilities of approach.

Figure 8
Line graph titled “Validation Loss Curve on eICU” showing validation loss versus epoch for six models. Each model is represented by a line in the legend: Proposed model, Khader et al. (2023), Kumar et al. (2024, 2025), Xie et al. (2023), and Narmadha et al. (2025). All curves generally decrease, indicating improvement in validation loss over fifty epochs.

Figure 8. Validation loss curve on the eICU dataset, indicating effective model generalization and minimal overfitting.

Figure 9
Line graph titled “Validation Loss Curve on Chest X-ray Images” showing validation loss decreasing over 50 epochs. The graph includes six lines representing different models: Proposed model, Khader et al. (2023), Kumar et al. (2024), Kumar et al. (2025), Xie et al. (2023), and Narmadha et al. (2025). Each line follows a similar downward trend, indicating improved model performance over epochs.

Figure 9. Validation loss trajectory on the Chest X-ray dataset, highlighting the model's strong generalization performance.

Figure 10
ROC curve comparing different models on MIMIC-IV dataset. Proposed model, Khader et al. (2023), Kumar et al. (2024, 2025), Xie et al. (2023), and Narmadha et al. (2025) with AUC values ranging from 0.47 to 0.50. A dashed line represents random guess. True and false positive rates range from 0 to 1.

Figure 10. ROC analysis for MIMIC-IV, where the proposed framework achieves an AUC exceeding 0.95, highlighting its strong discriminative power.

Figure 11
ROC curve graph for eICU dataset shows true positive rate versus false positive rate for multiple models. The models include Proposed Model (AUC 0.50), Khader et al. (2023, AUC 0.49), Kumar et al. (2024, AUC 0.47), Kumar et al. (2025, AUC 0.49), Xie et al. (2023, AUC 0.49), and Narmadha et al. (2025, AUC 0.48). A dashed line represents random guess.

Figure 11. ROC curve for the eICU dataset, where proposed method attains an AUC near 0.96, reflecting high diagnostic accuracy.

Figure 12
ROC curve graph showing the performance of various models on chest X-ray images. The x-axis represents the false positive rate, and the y-axis shows the true positive rate. Six models are compared, each with a different line color and area under the curve (AUC) values ranging from approximately 0.47 to 0.50. A dashed line represents random guessing.

Figure 12. ROC curve on the Chest X-ray dataset, where the proposed method's AUC near 0.96 signifies excellent discriminatory capability.

On the MIMIC-IV dataset, proposed method achieved an AUC of 0.97, while Khader et al. (2023), Kumar and Sharma (2024), and Kumar et al. (2025) recorded AUCs of 0.93, 0.91, and 0.94, respectively. The eICU and Chest X-ray datasets yielded comparable results, with the method consistently achieving AUCs above 0.95. These statistically significant improvements in AUC underscore the model's ability to reliably distinguish between positive and negative cases in a clinical setting.

7.2 Comparative analysis of state-of-the-art methods

The proposed approach was benchmarked against five state-of-the-art methods: Khader et al. (2023), Kumar and Sharma (2024), Kumar et al. (2025), Xie et al. (2023), and Narmadha and Gobalakrishnan (2025). Composite evaluation metrics are presented in Figures 1315. For instance, on the MIMIC-IV dataset, the model achieved an accuracy of ~ 93%, a sensitivity of 91%, and a specificity of 94%, which are 2–5% higher than the corresponding metrics of the baseline models. Detailed subplots (Figures 1618) further demonstrate that the proposed approach consistently outperforms in terms of F1 score and MCC, indicating robust performance even under conditions of class imbalance.

Figure 13
Bar chart comparing evaluation metrics across several models on MIMIC-IV. Metrics include accuracy, sensitivity, specificity, AUC, precision, F1 score, MCC, and Brier score, with the proposed model, Khader et al. (2023), Kumar et al. (2024 and 2025), Xie et al. (2023), and Namathia et al. (2025) displayed. Each model's metrics are color-coded accordingly.

Figure 13. Bar chart summarizing key evaluation metrics on the MIMIC-IV dataset, with proposed method outperforming state-of-the-art alternatives.

Figure 14
Bar chart titled “Evaluation Metrics Comparison on eICU” showing various models compared on metrics like Accuracy, Sensitivity, Specificity, AUC, Precision, F1 Score, MCC, and Brier Score. The proposed model shows higher metric values overall compared to models by Khader et al. (2023), Kumar et al. (2024), Kumar et al. (2025), Xie et al. (2023), and Namudha et al. (2025).

Figure 14. Grouped bar chart comparing evaluation metrics on the eICU dataset, with proposed model consistently leading in performance.

Figure 15
Bar chart comparing evaluation metrics for different models on chest X-ray images. The metrics include accuracy, sensitivity, specificity, AUC, precision, F1 score, MCC, and Brier score. Each model, including the proposed model and those by Khader et al., Kumar et al., Xie et al., and Namamba et al., displays a similar high performance across all metrics.

Figure 15. Bar chart of evaluation metrics for the Chest X-ray dataset, illustrating that the proposed model outperforms other approaches on multiple metrics.

Figure 16
Bar charts comparing various evaluation metrics for different models on MIMIC-IV. Metrics include Accuracy, Sensitivity, Specificity, AUC, Precision, F1 Score, MCC, and Brier Score. Each chart shows results for the proposed model and models by Khader et al. (2023), Kumar et al. (2024), Xie et al. (2025), and Narmadha et al. (2025). Performance is consistently depicted across different metrics.

Figure 16. Detailed performance evaluation on the MIMIC-IV dataset across multiple criteria.

Figure 17
Bar charts displaying evaluation metrics on eICU for various models, including Accuracy, Sensitivity, Specificity, AUC, Precision, F1 Score, MCC, and Brier Score. The proposed model and others like Khader et al., Kumar et al., Xie et al., and Namamba et al. are compared. Each chart shows similar performance across models within each metric.

Figure 17. A breakdown of detailed metrics on the eICU dataset.

Figure 18
Bar charts displaying evaluation metrics for chest X-ray images. Metrics include Accuracy, Sensitivity, Specificity, AUC, Precision, F1 Score, MCC, and Brier Score. Each chart compares a proposed model against models from Khader, Kumar, Xie, and Nwammaha between 2021 and 2023. Bars in all charts are relatively similar, except for Brier Score, where the proposed model performs better.

Figure 18. Detailed metric evaluation for the Chest X-ray dataset, which shows the nuanced differences in performance across the evaluated methods.

In addition to performance metrics, Table 3 presents a comprehensive overview of the computational efficiency and clinical responsiveness of each method. The proposed framework demonstrates an inference time of 25 ms and a memory footprint of 150 MB. In contrast, the baseline methods exhibit inference times ranging from 28 to 35 ms and memory footprints between 155 and 170 MB. Notably, the clinical responsiveness score, an aggregate measure of the system's ability to promptly trigger alerts, is highest for the proposed method (9.2 out of 10), highlighting its potential utility in real-time clinical environments.

Table 3
www.frontiersin.org

Table 3. Computational efficiency and clinical responsiveness metrics.

7.3 Clinical utility and intervention outcomes

The evaluation was extended beyond traditional performance metrics to examine the clinical utility of the proposed framework. In ICUs, reliable real-time monitoring and early warning systems are essential. The calibration curves (Figures 1921) demonstrate that the probability estimates of the proposed model are well-calibrated. The Brier score for this method consistently ranges between 0.07 and 0.09, in contrast to 0.09 to 0.12 for alternative approaches, indicating that the predictions are not only accurate but also reliable for clinical decision-making.

Figure 19
Calibration curve for MIMIC-IV dataset showing various models: Proposed, Khader et al. (2023), Kumar et al. (2024, 2025), Xie et al. (2023), and Narmadha et al. (2025). The y-axis represents the fraction of positives, and the x-axis denotes mean predicted probability. All models perform closely to the ideal calibration line.

Figure 19. Calibration curve for the MIMIC-IV dataset showing that the predicted probabilities align closely with observed outcomes.

Figure 20
Calibration curve graph on eICU data, plotting mean predicted probability on the x-axis against fraction of positives on the y-axis. It includes curves for models by various authors and an ideal calibration line.

Figure 20. Calibration plot for the eICU dataset, confirming the accuracy of predicted probabilities through close adherence to the ideal calibration line.

Figure 21
Calibration curve on chest X-ray images showing the fraction of positives versus mean predicted probability. Multiple models are plotted: the proposed model, Khader et al. (2023), Kumar et al. (2024), Kumar et al. (2025), Xie et al. (2023), Narmadha et al. (2025), along with the ideal calibration line. All models closely follow the ideal calibration line, indicating good calibration.

Figure 21. Calibration curve for the Chest X-ray dataset, which shows that the proposed model's predicted probabilities are highly consistent with observed outcomes.

In addition, the study examined the effects of personalized, AI-generated intervention plans on clinical outcomes. Table 4 presents a summary of key clinical outcome measures: the proposed method achieved a diaphragm atrophy reduction of 15.3%, a weaning success rate of 87.6%, and an early warning accuracy of 92.1%. These results demonstrate significant improvements compared to the average performance of current state-of-the-art methods, which reported a diaphragm atrophy reduction of 10.8%, a weaning success rate of 81.4%, and an early warning accuracy of 88.7%. The enhanced performance in these clinical outcomes is attributed to the advanced feature extraction and temporal modeling capabilities of the framework, which enable timely ventilator adjustments and personalized interventions.

Table 4
www.frontiersin.org

Table 4. Clinical outcomes and intervention efficacy.

7.4 Evaluation of dynamic ventilator parameter adjustment

The dynamic adjustment of ventilator parameters is essential for optimal patient management. The model's capability to predict trends in real-time enables clinicians to proactively modify ventilator settings. As demonstrated by the composite metrics in Figures 1315, there exists a strong correlation between the model's predictions and the actual ventilator settings observed in practice. This strong correlation is further substantiated by the low Brier scores reported in the calibration analyses.

The framework's dynamic adjustment capability effectively mitigates ventilator-induced lung injury and improves patient outcomes by ensuring ventilation is customized to the evolving needs of each patient. This is evidenced by the increased rates of successful weaning and the reduction in diaphragm atrophy observed in clinical outcomes. The integration of temporal predictions with real-time data establishes a seamless and continuous feedback loop, which is essential for expeditious clinical decision-making.

7.5 Feature importance and model interpretability

To enhance transparency and foster clinician trust, SHAP (SHapley Additive exPlanations) was employed to quantify the contribution of each input feature to the model's predictions. Figure 22 presents a summary beeswarm plot of the top 10 features ranked by their mean absolute SHAP values. In this plot, each point represents a single sample's SHAP value for that feature, with the horizontal spread indicating the variability of its impact. Notably, “Vent_Vol” and several imaging embedding dimensions (e.g., “Imb_3”) exhibit the largest positive and negative contributions, confirming their critical role in predicting diaphragm dysfunction.

Figure 22
Scatter plot showing SHAP values for the top ten feature importances. Features, including Phys_4, Imb_2, and Heart_Rate, are listed on the Y-axis. SHAP values range from -1 to 2 on the X-axis, indicating the impact of each feature on the model's output.

Figure 22. The SHAP summary beeswarm plot illustrates the distribution of feature contributions across all samples, emphasizing the top 10 most influential variables in the model's predictions.

To further examine how individual feature values influence the model output, Figure 23 displays a dependence plot for “Vent_Vol.” As ventilator tidal volume increases beyond ~ 550 ml, SHAP values rise sharply, indicating a higher predicted risk of dysfunction. The red lowess curve highlights this non-linear relationship, demonstrating that extreme ventilator settings disproportionately drive the prediction score. Collectively, these explainability analyses reveal both global feature importance and local decision behavior, supporting more informed clinical interpretation of model recommendations.

Figure 23
Scatter plot depicting SHAP values for ventilator volume, with a red trend line. The x-axis represents ventilator volume in milliliters (350 to 700), and the y-axis shows SHAP values ranging from -1.5 to 1.5. The data points are scattered around the trend line, indicating variable influence on the SHAP values.

Figure 23. The SHAP dependence plot for ventilator tidal volume demonstrates the non-linear increase in predicted dysfunction risk as tidal volume surpasses clinical thresholds.

7.6 Discussion

Table 5 presents a detailed analysis of accuracy, precision, recall, F1-score, and MCC for each method across different datasets. The proposed model consistently surpasses the baseline methods, achieving the highest accuracy rates of 92.3% on MIMIC-IV, 91.8% on eICU, and 92.0% on Chest X-ray, while also leading in precision, recall, and F1-score across all datasets. Additionally, its MCC values (0.84, 0.82, and 0.83, respectively) demonstrate a strong overall agreement, even in the presence of class imbalance. In contrast, other state-of-the-art methods exhibit lower performance, with MCCs ranging from 0.70 to 0.80. These findings highlight the robustness and balanced predictive capability of the proposed framework compared to existing approaches.

Table 5
www.frontiersin.org

Table 5. Detailed performance metrics across methods and datasets.

7.6.1 Comparative evaluation protocol

To ensure a fair comparison, all baseline methods Khader et al. (2023), Kumar and Sharma (2024), Kumar et al. (2025) were re-implemented and trained using the identical integrated dataset described in Section 3. The same preprocessing pipeline, which includes normalization, temporal alignment, and train/validation splits, as well as evaluation criteria, was applied to each model. Hyperparameters for each baseline were optimized via grid search under the same constraints (learning rate, batch size, number of epochs) used for the proposed model. Consequently, the performance figures reported in Table 5 reflect each method's results on the exact same data inputs and evaluation protocol, thereby eliminating any dataset-related bias in the comparative analysis.

7.6.2 Limitations

The proposed work also posses several limitations which can be considered to overcome in future studies. For example, although the integrated dataset draws from multiple publicly available sources, it remains retrospective and may not capture all relevant variations in patient populations or ventilator management protocols. Consequently, the model's generalizability to other institutions or prospective clinical workflows may be limited without further validation in multi-center studies.

Secondly, the proposed framework relies on synchronized multimodal inputs (imaging, physiological signals, ventilator data) that necessitate robust data acquisition and preprocessing pipelines. In real-world settings, missing or noisy data may degrade performance, and the impact of data quality on prediction accuracy has not been systematically evaluated. Future work should incorporate data-imputation strategies and uncertainty quantification to mitigate these effects. Thirdly, although SHAP-based explainability analyses were introduced, the complexity of Transformer architectures and GRU networks can still pose challenges for clinical interpretability. The SHAP explanations provide insight into feature importance, but they do not fully resolve the “black box” nature of deep learning models. Complementary methods, such as concept bottleneck models or prototype-based explanations, may be required to achieve greater transparency. Fourthly, the current evaluation focuses on binary prediction of diaphragm dysfunction. In practice, diaphragm performance spans a continuum of function and may benefit from regression-based or multi-class classification approaches that distinguish between mild, moderate, and severe impairment. Extending the framework to accommodate graded outcomes could enhance clinical utility.

Finally, the computational demands of the hierarchical Transformer and temporal modules may limit deployment on resource-constrained devices. While inference times were acceptable on modern hardware, optimizations, such as model pruning, quantization, or knowledge distillation, will be necessary to support real-time implementation in diverse ICU environments.

8 Conclusion

In this study, an innovative AI-driven framework is presented, that leverages transformer-based deep learning and multimodal data integration to address diaphragm dysfunction in elderly patients undergoing mechanical ventilation. This study integrates heterogeneous clinical data, encompassing imaging, physiological measurements, and ventilator parameters, through a hierarchical Transformer encoder, attention-guided cross-modal fusion, and a temporal network to achieve superior predictive performance characterized by high AUC, robust calibration, and enhanced real-time monitoring.

An extensive evaluation has demonstrated that the proposed AI-driven framework significantly outperforms state-of-the-art methods in predicting diaphragm dysfunction across three diverse clinical datasets, achieving AUC values of 0.97, 0.96, and 0.96 on MIMIC-IV, eICU, and Chest X-ray data, respectively, with an average accuracy exceeding 92%. Moreover, clinical-outcome simulations indicate a 15.3% reduction in diaphragm atrophy (vs. 10.8% for competing approaches) and an 87.6% weaning success rate (compared to 81.4%), underscoring its potential for improving patient care. SHAP-based explainability further confirmed that ventilator volume and specific imaging embeddings are the most influential predictors, enhancing trust in automated recommendations.

Despite these advancements, several research directions remain to be explored: prospective, multicenter validation is necessary to confirm generalizability; the integration of regression and multi-class outputs could facilitate graded assessments of diaphragm health; and deployment strategies such as model pruning or knowledge distillation should be investigated to support real-time, resource-constrained environments. By addressing these areas, the framework can be refined for seamless clinical adoption, ultimately contributing to more proactive, personalized ventilation strategies in critical care.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

MX: Writing – original draft, Formal analysis, Writing – review & editing, Data curation, Methodology, Conceptualization, Investigation. ZJ: Writing – original draft, Methodology, Investigation, Resources, Funding acquisition. YM: Visualization, Writing – review & editing, Conceptualization, Investigation, Formal analysis, Data curation. ZY: Investigation, Software, Writing – review & editing, Methodology, Supervision, Data curation. LF: Writing – original draft, Formal analysis, Resources, Conceptualization, Data curation. JJ: Investigation, Validation, Project administration, Formal analysis, Writing – review & editing, Data curation, Methodology. DL: Funding acquisition, Writing – review & editing, Investigation, Supervision, Writing – original draft, Resources, Validation, Formal analysis, Visualization.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by the Jilin Provincial Science and Technology Development Program (20250203012SF), Science and Technology Research Project of the Jilin Provincial Department of Education (JJKH20250212KJ), and Jilin Provincial Medical and Health Talent Special Program (2024WSZX-B08).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Adibi, A., Cao, X., Ji, Z., Kaur, J. N., Chen, W., Healey, E., et al. (2025). Recent advances, applications and open challenges in machine learning for health: Reflections from research roundtables at ml4h 2024 symposium. arXiv preprint arXiv:2502.06693. doi: 10.48550/arXiv.2502.06693

Crossref Full Text | Google Scholar

Ahmed, F. R., Al-Yateem, N., Nejadghaderi, S. A., Saifan, A. R., Abdelaliem, S. M. F., and AbuRuz, M. E. (2025). Harnessing machine learning for predicting successful weaning from mechanical ventilation: a systematic review. Aust. Crit. Care 38:101203. doi: 10.1016/j.aucc.2025.101203

PubMed Abstract | Crossref Full Text | Google Scholar

AlSaad, R., Abd-Alrazaq, A., Boughorbel, S., Ahmed, A., Renault, M.-A., Damseh, R., et al. (2024). Multimodal large language models in health care: applications, challenges, and future outlook. J. Med. Internet Res. 26:e59505. doi: 10.2196/59505

PubMed Abstract | Crossref Full Text | Google Scholar

Bacanin, N., Jovanovic, L., Stoean, R., Stoean, C., Zivkovic, M., Antonijevic, M., et al. (2024a). Respiratory condition detection using audio analysis and convolutional neural networks optimized by modified metaheuristics. Axioms 13:335. doi: 10.3390/axioms13050335

Crossref Full Text | Google Scholar

Bacanin, N., Jovanovic, L., Zivkovic, M., Zivkovic, T., Bisevac, P., Dobrojevic, M., et al. (2024b). “Parkinson's disease detection with deep long short-term memory networks optimized by modified metaheuristic algorithm,” in Collective Intelligence (CRC Press), 204–229. doi: 10.1201/9781032690711-9

Crossref Full Text | Google Scholar

Battaglini, D., and Rocco, P. R. (2024). Challenges in transitioning from controlled to assisted ventilation in acute respiratory distress syndrome (ARDS) management. J. Clin. Med. 13:7333. doi: 10.3390/jcm13237333

PubMed Abstract | Crossref Full Text | Google Scholar

Bhakar, S., Sinwar, D., Pradhan, N., Dhaka, V. S., Cherrez-Ojeda, I., Parveen, A., et al. (2023). Computational intelligence-based disease severity identification: a review of multidisciplinary domains. Diagnostics 13:1212. doi: 10.3390/diagnostics13071212

PubMed Abstract | Crossref Full Text | Google Scholar

Biesheuvel, L. A., Dongelmans, D. A., and Elbers, P. W. (2024). Artificial intelligence to advance acute and intensive care medicine. Curr. Opin. Crit. Care 30, 246–250. doi: 10.1097/MCC.0000000000001150

PubMed Abstract | Crossref Full Text | Google Scholar

Duan, J., Xiong, J., Li, Y., and Ding, W. (2024). Deep learning based multimodal biomedical data fusion: an overview and comparative review. Inf. Fusion 112:102536. doi: 10.1016/j.inffus.2024.102536

Crossref Full Text | Google Scholar

Ganatra, H. A., and Varisco, B. M. (2019). “Respiratory complications of intensive care,” in Kendig's Disorders of the Respiratory Tract in Children (Elsevier), 596–605. doi: 10.1016/B978-0-323-44887-1.00037-7

Crossref Full Text | Google Scholar

Ganiya, R. K., Veeraiah, D., Thatha, V. N., Rao, K. S., Rao, J. N., Manjith, R., et al. (2024). Revolutionizing vascular health through the temporal convolutional transformer for drug screening and model evolution. Biomed. Signal Process. Control 95:106390. doi: 10.1016/j.bspc.2024.106390

Crossref Full Text | Google Scholar

Hantos, Z. (2021). Intra-breath oscillometry for assessing respiratory outcomes. Curr. Opin. Physiol. 22:100441. doi: 10.1016/j.cophys.2021.05.004

Crossref Full Text | Google Scholar

Howroyd, F., Chacko, C., MacDuff, A., Gautam, N., Pouchet, B., Tunnicliffe, B., et al. (2024). Ventilator-associated pneumonia: pathobiological heterogeneity and diagnostic challenges. Nat. Commun. 15:6447. doi: 10.1038/s41467-024-50805-z

PubMed Abstract | Crossref Full Text | Google Scholar

Hsia, C. C., Bates, J. H., Driehuys, B., Fain, S. B., Goldin, J. G., Hoffman, E. A., et al. (2023). Quantitative imaging metrics for the assessment of pulmonary pathophysiology: an official american thoracic society and fleischner society joint workshop report. Ann. Am. Thorac. Soc. 20, 161–195. doi: 10.1513/AnnalsATS.202211-915ST

PubMed Abstract | Crossref Full Text | Google Scholar

Jiang, J., Chen, L., Ke, L., Dou, B., Zhang, C., Feng, H., et al. (2024). A review of transformers in drug discovery and beyond. J. Pharm. Anal. 15:101081. doi: 10.1016/j.jpha.2024.101081

PubMed Abstract | Crossref Full Text | Google Scholar

Johnson, A., Bulgarelli, L., Pollard, T., Horng, S., Celi, L. A., and Mark, R. (2023). Mimic-iv Clinical Database Demo (version 2.2). Atlanta, GA: PhysioNet.

Google Scholar

Jovanovic, L., Bacanin, N., Zivkovic, M., Antonijevic, M., Petrovic, A., and Zivkovic, T. (2023). “Anomaly detection in ECG using recurrent networks optimized by modified metaheuristic algorithm,” in 2023 31st telecommunications forum (TELFOR) (Belgrade: IEEE), 1–4. doi: 10.1109/TELFOR59449.2023.10372802

Crossref Full Text | Google Scholar

Jung, B., Moury, P. H., Mahul, M., de Jong, A., Galia, F., Prades, A., et al. (2016). Diaphragmatic dysfunction in patients with icu-acquired weakness and its impact on extubation failure. Intensive Care Med. 42, 853–861. doi: 10.1007/s00134-015-4125-2

PubMed Abstract | Crossref Full Text | Google Scholar

Kaya, P. K., and Iscimen, R. (2025). Management of mechanical ventilation and weaning in critically ill patients with neuromuscular disorders. Respir. Med. 237:107951. doi: 10.1016/j.rmed.2025.107951

PubMed Abstract | Crossref Full Text | Google Scholar

Kermany, D., Zhang, K., and Goldbaum, M. (2018). “Labeled optical coherence tomography (OCT) and chest x-ray images for classification.” Mendeley data.

Google Scholar

Khader, F., Müller-Franzes, G., Wang, T., Han, T., Tayebi Arasteh, S., Haarburger, C., et al. (2023). Multimodal deep learning for integrating chest radiographs and clinical parameters: a case for transformers. Radiology 309:e230806. doi: 10.1148/radiol.230806

PubMed Abstract | Crossref Full Text | Google Scholar

Kim, J., Kim, G.-H., Kim, J.-W., Kim, K. H., Maeng, J.-Y., Shin, Y.-G., et al. (2025). Transformer-based model for predicting length of stay in intensive care unit in sepsis patients. Front. Med. 11:1473533. doi: 10.3389/fmed.2024.1473533

PubMed Abstract | Crossref Full Text | Google Scholar

Kumar, S., and Sharma, S. (2024). An improved deep learning framework for multimodal medical data analysis. Big Data Cogn. Comput. 8:125. doi: 10.3390/bdcc8100125

Crossref Full Text | Google Scholar

Kumar, S., Sharma, S., and Megra, K. T. (2025). Transformer enabled multi-modal medical diagnosis for tuberculosis classification. J. Big Data 12:5. doi: 10.1186/s40537-024-01054-w

Crossref Full Text | Google Scholar

Kumar, S., and Solanki, A. (2023). An abstractive text summarization technique using transformer model with self-attention mechanism. Neural Comput. Appl. 35, 18603–18622. doi: 10.1007/s00521-023-08687-7

Crossref Full Text | Google Scholar

Laghi, F. A., Saad, M., and Shaikh, H. (2021). Ultrasound and non-ultrasound imaging techniques in the assessment of diaphragmatic dysfunction. BMC Pulm. Med. 21, 1–29. doi: 10.1186/s12890-021-01441-6

PubMed Abstract | Crossref Full Text | Google Scholar

Lessa, T. B., de Abreu, D. K., Bertassoli, B. M., and Ambrósio, C. E. (2016). Diaphragm: A vital respiratory muscle in mammals. Ann. Anat. 205, 122–127. doi: 10.1016/j.aanat.2016.03.008

PubMed Abstract | Crossref Full Text | Google Scholar

Liao, K.-M., Liu, C.-F., Chen, C.-J., and Shen, Y.-T. (2021). Machine learning approaches for predicting acute respiratory failure, ventilator dependence, and mortality in chronic obstructive pulmonary disease. Diagnostics 11:2396. doi: 10.3390/diagnostics11122396

PubMed Abstract | Crossref Full Text | Google Scholar

Llamas-Álvarez, A. M., Tenza-Lozano, E. M., and Latour-Pérez, J. (2017). Diaphragm and lung ultrasound to predict weaning outcome: systematic review and meta-analysis. Chest 152, 1140–1150. doi: 10.1016/j.chest.2017.08.028

PubMed Abstract | Crossref Full Text | Google Scholar

Ma, X., Zhang, P., Zhang, S., Duan, N., Hou, Y., Zhou, M., et al. (2019a). A tensorized transformer for language modeling. Adv. Neural Inf. Process. Syst. 32, 1–11. doi: 10.48550/arXiv.1906.09777

Crossref Full Text | Google Scholar

Ma, Y., Zhang, J., Niu, D., Hassan, M. U., and Zhao, X. (2019b). “An unsupervised approach for 3D medical image registration,” in Proceedings of the 2019 7th International Conference on Information Technology: IoT and Smart City, 259–263. doi: 10.1145/3377170.3377273

Crossref Full Text | Google Scholar

Marey, A., Arjmand, P., Alerab, A. D. S., Eslami, M. J., Saad, A. M., Sanchez, N., et al. (2024). Explainability, transparency and black box challenges of ai in radiology: impact on patient care in cardiovascular radiology. Egypt J. Radiol. Nucl. Med. 55:183. doi: 10.1186/s43055-024-01356-2

Crossref Full Text | Google Scholar

Markovic, F., Jovanovic, L., Spalevic, P., Kaljevic, J., Zivkovic, M., Simic, V., et al. (2025). Parkinsons detection from gait time series classification using modified metaheuristic optimized long short term memory. Neural Process. Lett. 57:14. doi: 10.1007/s11063-025-11735-z

Crossref Full Text | Google Scholar

Mostafaei, S. H., Tanha, J., and Sharafkhaneh, A. (2024). A novel deep learning model based on transformer and cross modality attention for classification of sleep stages. J. Biomed. Inform. 157:104689. doi: 10.1016/j.jbi.2024.104689

PubMed Abstract | Crossref Full Text | Google Scholar

Narmadha, A. P., and Gobalakrishnan, N. (2025). Het-rl: Multiple pulmonary disease diagnosis via hybrid efficient transformers based representation learning model using multi-modality data. Biomed. Signal Process. Control 100:107157. doi: 10.1016/j.bspc.2024.107157

Crossref Full Text | Google Scholar

Panelli, A., Grunow, J. J., Verfu, M. A., Bartels, H. G., Brass, Z., and Schaller, S. J. (2024). Outcomes in critically ill patients after diaphragmatic stimulation on ventilator-induced diaphragmatic dysfunction: a systematic review. Eur. J. Phys. Rehabil. Med. 59:772. doi: 10.23736/S1973-9087.23.08031-0

PubMed Abstract | Crossref Full Text | Google Scholar

Pollard, T. J., Johnson, A. E. W., Raffa, J. D., Celi, L. A., Mark, R. G., and Badawi, O. (2018). The eicu collaborative research database, a freely available multi-center database for critical care research. Sci. Data 5:180178. doi: 10.1038/sdata.2018.178

PubMed Abstract | Crossref Full Text | Google Scholar

Powers, S. K. (2024). Ventilator-induced diaphragm dysfunction: phenomenology and mechanism (s) of pathogenesis. J. Physiol. 602, 4729–4752. doi: 10.1113/JP283860

PubMed Abstract | Crossref Full Text | Google Scholar

Powers, S. K., Kavazis, A. N., and Levine, S. (2009). Prolonged mechanical ventilation alters diaphragmatic structure and function. Crit. Care Med. 37, S347–S353. doi: 10.1097/CCM.0b013e3181b6e760

PubMed Abstract | Crossref Full Text | Google Scholar

Purkovic, S., Jovanovic, L., Zivkovic, M., Antonijevic, M., Dolicanin, E., Tuba, E., et al. (2024). Audio analysis with convolutional neural networks and boosting algorithms tuned by metaheuristics for respiratory condition classification. J. King Saud Univ. Comput. Inf. Sci. 36:102261. doi: 10.1016/j.jksuci.2024.102261

Crossref Full Text | Google Scholar

Rashno, E., Eskandari, A., Anand, A., and Zulkernine, F. (2024). Survey: Transformer-based models in data modality conversion. arXiv preprint arXiv:2408.04723. doi: 10.48550/arXiv.2408.04723

Crossref Full Text | Google Scholar

Ratano, D., Zhang, B., Dianti, J., Georgopoulos, D., Brochard, L. J., Chan, T. C., et al. (2024). Lung-and diaphragm-protective strategies in acute respiratory failure: an in silico trial. Intensive Care Med. Exp. 12:20. doi: 10.1186/s40635-024-00606-x

PubMed Abstract | Crossref Full Text | Google Scholar

Sarwar, R., Teh, P. S., Sabah, F., Nawaz, R., Hameed, I. A., Hassan, M. U., et al. (2024). Agi-p: A gender identification framework for authorship analysis using customized fine-tuning of multilingual language model. IEEE Access 12, 15399–15409. doi: 10.1109/ACCESS.2024.3358199

Crossref Full Text | Google Scholar

Siniscalchi, C., Nouvenne, A., Cerundolo, N., Meschi, T., Ticinesi, A., and in Emergency-Urgency Medicine Interest Group on Thoracic Ultrasound, P. P.-G. S. S. (2024). Diaphragm ultrasound in different clinical scenarios: a review with a focus on older patients. Geriatrics 9:70. doi: 10.3390/geriatrics9030070

PubMed Abstract | Crossref Full Text | Google Scholar

Torrini, F., Gendreau, S., Morel, J., Carteaux, G., Thille, A. W., Antonelli, M., et al. (2021). Prediction of extubation outcome in critically ill patients: a systematic review and meta-analysis. Critical Care 25, 1–12. doi: 10.1186/s13054-021-03802-3

PubMed Abstract | Crossref Full Text | Google Scholar

Vassilakopoulos, T., and Petrof, B. J. (2004). Ventilator-induced diaphragmatic dysfunction. Am. J. Respir. Crit. Care Med. 169, 336–341. doi: 10.1164/rccm.200304-489CP

PubMed Abstract | Crossref Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). Attention is all you need. Adv. Neural Inf. Process. Syst. 30, 1–15. doi: 10.48550/arXiv.1706.03762

Crossref Full Text | Google Scholar

Xie, W., Fang, Y., Yang, G., Yu, K., and Li, W. (2023). Transformer-based multi-modal data fusion method for copd classification and physiological and biochemical indicators identification. Biomolecules 13:1391. doi: 10.3390/biom13091391

PubMed Abstract | Crossref Full Text | Google Scholar

Xu, P., Zhu, X., and Clifton, D. A. (2023). Multimodal learning with transformers: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 45, 12113–12132. doi: 10.1109/TPAMI.2023.3275156

PubMed Abstract | Crossref Full Text | Google Scholar

Xu, X., Li, J., Zhu, Z., Zhao, L., Wang, H., Song, C., et al. (2024). A comprehensive review on synergy of multi-modal data and ai technologies in medical diagnosis. Bioengineering 11:219. doi: 10.3390/bioengineering11030219

PubMed Abstract | Crossref Full Text | Google Scholar

Yamazaki, L. (2024). Investigation of Transformers and Other Machine Learning Techniques for Health Data Classification (PhD thesis). Murfreesboro, TN: Middle Tennessee State University.

Google Scholar

Yang, J., Hao, S., Huang, J., Chen, T., Liu, R., Zhang, P., et al. (2023). The application of artificial intelligence in the management of sepsis. Med. Rev. 3, 369–380. doi: 10.1515/mr-2023-0039

PubMed Abstract | Crossref Full Text | Google Scholar

Yayan, J., and Schiffner, R. (2024). Weaning failure in elderly patients: a systematic review and meta-analysis. J. Clin. Med. 13:6429. doi: 10.3390/jcm13216429

PubMed Abstract | Crossref Full Text | Google Scholar

Yoon, J. H., Pinsky, M. R., and Clermont, G. (2022). Artificial intelligence in critical care medicine. Annu. Update Intensive Care Emerg. Med. 2022, 353–367. doi: 10.1007/978-3-030-93433-0_27

Crossref Full Text | Google Scholar

Zhang, J., Contreras, M., Bandyopadhyay, S., Davidson, A., Sena, J., Ren, Y., et al. (2024). Mango: Multimodal acuity transformer for intelligent icu outcomes. arXiv preprint arXiv:2412.17832. doi: 10.48550/arXiv.2412.17832

Crossref Full Text | Google Scholar

Zhang, M., Hassan, M. U., Niu, D., Zhao, X., Nawaz, R., Hameed, I. A., et al. (2023). A methodology for shape matching of non-rigid structures based on integrated graphical information. Displays 76:102336. doi: 10.1016/j.displa.2022.102336

Crossref Full Text | Google Scholar

Zivkovic, M., Bacanin, N., Antonijevic, M., Nikolic, B., Kvascev, G., Marjanovic, M., et al. (2022). Hybrid cnn and xgboost model tuned by modified arithmetic optimization algorithm for covid-19 early diagnostics from x-ray images. Electronics 11:3798. doi: 10.3390/electronics11223798

Crossref Full Text | Google Scholar

Keywords: transformer models, multimodal data integration, diaphragm dysfunction, mechanical ventilation, predictive modeling, critical care applications

Citation: Xinli M, Jie Z, Ming Y, Yanping Z, Fan L, Jing J and Lu D (2025) Transformer-based multimodal precision intervention model for enhancing diaphragm function in elderly patients. Front. Comput. Neurosci. 19:1615576. doi: 10.3389/fncom.2025.1615576

Received: 21 April 2025; Accepted: 18 July 2025;
Published: 18 August 2025.

Edited by:

Milos Antonijevic, Singidunum University, Serbia

Reviewed by:

Nebojsa Bacanin, Singidunum University, Serbia
Miodrag Zivkovic, Singidunum University, Serbia

Copyright © 2025 Xinli, Jie, Ming, Yanping, Fan, Jing and Lu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Ding Lu, ZGluZ2x1LmpsdUBvdXRsb29rLmNvbQ==; NzU1OTQyNjZAcXEuY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.