Skip to main content

ORIGINAL RESEARCH article

Front. Neurol., 19 December 2023
Sec. Stroke
This article is part of the Research Topic Artificial Intelligence in Acute Neurology View all 11 articles

Research on prognostic risk assessment model for acute ischemic stroke based on imaging and multidimensional data

  • 1Postgraduate Cultivation Base of Guangzhou University of Chinese Medicine, Panyu Central Hospital, Guangzhou, China
  • 2Graduate School, Guangzhou University of Chinese Medicine, Guangzhou, China
  • 3Medical Imaging Institute of Panyu, Guangzhou, China
  • 4Radiology Department of Sun Yat-sen Memorial Hospital, Sun Yat-sen University, Guangzhou, China
  • 5Laboratory for Intelligent Information Processing, Guangdong University of Technology, Guangzhou, China
  • 6Kerry Rehabilitation Medicine Research Institute, Shenzhen, China
  • 7Stroke Biological Recovery Laboratory, Spaulding Rehabilitation Hospital, Teaching Affiliate of Harvard Medical School, Charlestown, MA, United States
  • 8Panyu Health Management Center, Guangzhou, China

Accurately assessing the prognostic outcomes of patients with acute ischemic stroke and adjusting treatment plans in a timely manner for those with poor prognosis is crucial for intervening in modifiable risk factors. However, there is still controversy regarding the correlation between imaging-based predictions of complications in acute ischemic stroke. To address this, we developed a cross-modal attention module for integrating multidimensional data, including clinical information, imaging features, treatment plans, prognosis, and complications, to achieve complementary advantages. The fused features preserve magnetic resonance imaging (MRI) characteristics while supplementing clinical relevant information, providing a more comprehensive and informative basis for clinical diagnosis and treatment. The proposed framework based on multidimensional data for activity of daily living (ADL) scoring in patients with acute ischemic stroke demonstrates higher accuracy compared to other state-of-the-art network models, and ablation experiments confirm the effectiveness of each module in the framework.

1 Introduction

Stroke is a cerebrovascular disease characterized by localized cerebral ischemia, hypoxia leading to ischemic necrosis or softening, and subsequent neurological dysfunction. Approximately 16 million people worldwide suffer from stroke each year, with 5.7 million deaths and around 5 million disabilities (1). Survivors often experience difficulties in swallowing, speech impairment, motor dysfunction, cognitive impairment, emotional disorders, and other functional deficits (2, 3). Early diagnosis, prediction, and rehabilitation are key strategies for improving the prognosis of stroke patients. Stroke treatment guidelines emphasize that early diagnosis of stroke relies on imaging findings and clinical symptoms/signs. Neuroimaging plays a crucial role in the definitive diagnosis of suspected stroke patients. However, limited studies have been conducted on the correlation between imaging-based predictions of complications in acute ischemic stroke, and most of them focus on single complications (4, 5). Several scales have been used clinically to predict functional outcomes in stroke patients (6, 7), such as the acute stroke registry and analysis of Lausanne (ASTRAL) and the totaled health risks in vascular events (THRIVE). However, these scales mostly incorporate variables at admission and are intended to provide information for treatment, without collecting post-treatment data for prediction. Therefore, in order to accurately assess the prognostic outcomes of patients, adjust treatment plans in a timely manner for those with poor prognosis, and intervene in modifiable risk factors, machine learning methods are needed to predict the prognostic risk of patients with acute ischemic stroke.

MRI is one of the crucial tools for evaluating acute ischemic stroke and has been widely used in clinical practice due to its high detection accuracy, sensitivity, and specificity. Computer-aided diagnosis (CAD) based on MRI has received extensive attention from researchers both domestically and internationally. For example, the texture analysis of apparent diffusion coefficient maps and diffusion-weighted imaging were used to predict the prognosis and subtype of ischemic stroke (810). A systematic review also demonstrated that a combined model combining clinical and imaging variables was more predictive of stroke outcome (11).

This study aimed to construct an acute ischemic stroke prognosis assessment model based on multidimensional imaging data, clinical information, treatment plans, prognosis, and complications, as shown in Figure 1. It contains the adaptive lesion awareness module (ALAM), the patient metadata encoder based on multilayer perceptron (MLP-FE), the cross-modal attention module (CMAM). This model provided a scientific basis for early clinical intervention, enabling healthcare professionals to make informed decisions and interventions based on the predicted prognosis.

FIGURE 1
www.frontiersin.org

Figure 1. Based on the multi-dimensional network framework for the ADL score in patients with acute cerebral infarction.

2 Multidimensional ADL scoring network framework for acute ischemic stroke patients

In this section, we developed a multidimensional data-based cross-modal attention fusion network for the prognosis assessment of acute ischemic stroke patients. The structure of the network was illustrated in Figure 2 and consisted of three main components: the ALAM, the MLP-FE, and the CMAM, which were described as follows:

FIGURE 2
www.frontiersin.org

Figure 2. Example of the AIS patient MRI, where the red arrow indicates the focal area.

Firstly, the ALAM was based on a multi-scale global-local attention mechanism. It adaptively focused on the lesion region by learning joint features of global context information and local detailed features, enabling the extraction of more discriminative imaging features (shown as Figure 1).

Secondly, to fully utilize the patient metadata, we designed the MLP-FE. This encoder compiled the patient metadata into corresponding feature representations.

Lastly, the CMAM was proposed in this study for the fusion of multidimensional data, including clinical information, imaging features, treatment plans, prognosis, and complications. It aimed to achieve information complementarity and cross-validation by integrating data from different modalities.

2.1 Adaptive lesion sense module

In clinical practice, acute ischemic stroke patients exhibited complex MRI imaging features (shown in Figure 2): the location, size, and appearance of the lesions varied significantly among different patients and stages of development. Furthermore, the complications associated with stroke often had similar imaging characteristics, making them difficult to differentiate. Additionally, mild cases with smaller lesions were prone to being overlooked. This posed a significant challenge for deep learning models to learn lesion imaging features from MRI, as it involved the issues of “same disease, different images” and “different diseases, same images.”

In this subsection, we constructed an integration of the global-local attention module (GLAM) and the multiscale joint feature fusion module. This integration allowed for adaptive attention towards the relevant lesion regions and extraction of MRI images features by combining global and local views, akin to the process of radiologists’ image interpretation. Additionally, the hierarchical design within this module facilitated multiscale joint feature fusion, enabling the extraction of lesion characteristics of different shapes and sizes. Moreover, these features interacted with each other to obtain the most effective image information.

2.1.1 The global-local attention module

The global-local attention module is used to extract joint features that capture both global context and local details. It consists of two attention blocks: the global attention block (GAB) based on self-attention mechanism and the local attention block (LAB) based on channel spatial attention. The GAB learns global context information to provide a comprehensive understanding of the scene, enabling the localization of lesion regions and the suppression of irrelevant background information. The LAB refines local features to capture more detailed lesion information, which helps in distinguishing similar imaging features of complications in acute ischemic stroke (AIS) and addressing the issue of “different diseases, same images.”

Global context encompasses the implicit relationships between pixels and scene information in an image, providing a holistic perception of the scene. It has been widely used in various computer vision applications such as scene parsing and object detection (1215). Intuitively, not all image content in MRI contributes to the final diagnosis, and irrelevant background information may even have a detrimental effect. However, by reasoning about the global scene, it becomes easier to detect and focus on the lesion region in the image. Existing methods primarily rely on convolutional neural networks (CNNs) for extracting MRI images features, which do not fully exploit and utilize the global context of the image. This is mainly due to the local nature of CNNs, which prevents them from learning the global context that aids in better lesion localization. Currently, effective modeling and integration of global context information in MRI images feature learning is an important research question that has not been fully investigated.

Inspired by natural language processing (NLP) networks, which extensively use transformers (14) to model global dependencies in language sequences, we designed a global attention block based on self-attention Transformer to model non-local interactions for learning global contextual information from local features. This global attention mechanism effectively detects lesion regions and efficiently disregards irrelevant background information. It is worth noting that, unlike segmentation-based methods, the GAB can locate lesion regions in an adaptive learning manner without any manual annotation. This enables more flexibility and robustness in handling various complex lesion shapes. The implementation process can be described as follows.

Taking into account that GAB required a sequence as input, similar to (15), the input feature map FRC×H×W was transformed into a sequence of flattened 2D patches XpRP2C×L, where (P, P) represented the resolution of each image patch, L=HW/P2 was the number of image patches. Then, the image patches were mapped to a D-dimensional space through a learnable linear transformation, resulting in an output feature sequence XRD×L which served as the input to GAB.

As shown in Figure 3, GAB consisted of two sub-layers: Multi-head self-attention (MHSA) and multi-layer perceptron (MLP). Each sub-layer was surrounded by residual connections, and layer normalization (LN) was applied before MHSA and MLP. Therefore, for an input feature sequence X, the output of GAB was given by

Fg=X+MLP[LN(X)],    (1)

among,

X=X+MHSA[LN(X)].    (2)
FIGURE 3
www.frontiersin.org

Figure 3. The structure of global-local attention module. Global attention block: learning global context information for detecting lesion regions; local attention block: refining local features to learn lesion details.

Although GAB can detect lesion regions through globally-guided attention, it is limited in learning lesion details. As the complications of AIS have similar imaging features, that is, there is a situation of “heterogeneous lesions and simultaneous shadows,” which requires learning more detailed lesion features to distinguish them. However, traditional CNN is difficult to accurately extract subtle visual features from images (16). The features extracted by CNN contain two types of information: channels and spatial regions, which have different contributions to learning lesion details (17, 18). In order to improve the characteristics of network contribution to high sensitivity, make it concentrate more high-end features, in order to enhance expressing ability for the pathological change information. In this study, a LAB based on channel spatial attention was designed to conduct adaptive feature refinement and refinement by learning the relationship between channel and feature spatial location.

As shown in Figure 3, LAB was constructed sequentially from channels and spatial attention to effectively aid the flow of information within the network by learning which information to emphasize or suppress. Based on this, the LAB can be adaptive to refine local characteristic lesions in order to obtain more detailed information, and thus improve the network’s ability to distinguish the similarity between classes of AIS complications.

Given a figure FRC×H×W as input, LAB sequentially output a channel attention map ΑttcRC×1×1 and a 2D spatial attention map ΑttsR1×H×W, which can be defined as

F=Attc(F)FFl=Atts(F)F    (3)

In this context, represented the dot product, F and Fl correspond to the channel branch and LAB output, respectively.

2.1.2 Multi-scale joint feature fusion module

Although many deep learning-based methods have achieved great success in stroke lesion detection, most of them rely on single-scale feature information and cannot fully utilize multi-scale feature representation, so it is difficult to deal with lesions of various scales and shapes, especially for small lesions (19, 20). This is because the following several pooling layer resolution is reduced, the characteristics of the small lesions may be lost. Therefore, this project designs a network with multi-scale feature information to realize the comprehensive extraction of MRI lesion features. It can combine the convolutional features of lesions of different scales and sizes together, retain low-level features of small lesions in the feature extraction process, and improve the diagnostic performance of early and mild patients.

In ALAM, a multi-scale feature extractor based on the pre-trained VGG16 (21) was first used to extract the feature representation of different layers. Then, at each scale, GLAM took the extracted feature map as input and learns joint features, implemented by two attention branches, capturing both global context and local refinement features. Finally, by integrating GLAM into the feature extractor, ALAM was designed to propagate joint features in a coarse-to-fine manner, fully extracted useful explicit feature representations at different scales, and exploit the complementary advantages of cross-scale implicit correlations. Therefore, the proposed network can effectively and robustly learn the imaging features of complex lesions by fusing multi-scale joint features, thereby significantly improving the detection performance of stroke lesions.

2.2 Based on the MLP patient metadata encoder

In clinical practice, radiologists will make medical diagnosis and evaluation of patients based on their imaging features and multi-dimensional clinical information database, including clinical data, past medical history, complications, etc. However, existing deep learning (DL) models only use a single imaging feature to predict the diagnosis of patients, which is not convincing and accurate. In view of this, in addition to imaging features, this project also established a multi-dimensional clinical information database, including clinical data, past medical history, complications, and self-care scores, which can provide patients with higher quality, more accurate and more personalized medical diagnosis AIS.

In order to make full use of patient metadata, this study selected 19 metadata features from the multi-dimensional clinical information database of patients, and designed a patient metadata encoder based on MLP to compile the patient metadata into the corresponding feature representation. Nineteen metadata features included “hospital,” “gender,” “age “, “disorder “, “consciousness,” “complications (hemiplegia) “, “pneumonia,” “aphasia,” “swallowing disorder,” “facial paralysis,” “dementia,” “cognitive impairment,” “depression,” “first ability score,” “hypertension,” “diabetes,” “atrial fibrillation,” “coronary atherosclerotic heart disease,” “hyperlipidemia,” “high homocysteinemia.” Specifically, there were two dense layers in this encoder, including 256 neurons. Each layer was followed by batch normalization and rectified linear unit (ReLU) activation layers. After the first dense layer attempts to generalize the metadata into the network, a further 25% was discarded using the dropout layer, thus avoiding possible overfitting of the data.

2.3 Cross-modal attention module

In order to strengthen the transfer between features and improve the performance of multimodal features, this study proposed a cross-modal attention module for multi-dimensional data information fusion such as clinical data, imaging features, treatment plans, prognosis and complications to achieve information complementarity and cross-validation. Multimodal fusion features not only retained the imaging features of MRI, but also made up for the clinical relevant information of patients, which provided more rich and comprehensive information for clinical diagnosis and treatment. The specific implementation process was as follows:

Firstly, the imaging features F1 extracted by the adaptive lesion perception module and the metadata features F2 obtained by the patient metadata encoder were combined into a multimodal joint feature F:

F=Concat(F1,F2).    (4)

Then, it was considered that the operation of directly concatenating these two features might propagate a large amount of useless information and noise generated during the encoding process to the decoding layer. Therefore, this project proposed to model the relationship between features by cross-modal attention to generate attention masks, and used it to adaptively select important features, which could inhibit the spread of some harmful information to a certain extent. The attention mask A can be obtained from the following formula:

Α=σ{FC{ReLU{ΒN[FC(F)]}}},    (5)

in which, σ and ReLU were sigmoid and rectified linear unit activation functions, fully connected layer (FC) and batch normalization (BN) were fully connected layer and batch normalization, respectively.

Finally, the learned attention mask A was multiplied with the original feature map F to generate a feature map with attention weights Fo:

Fo=AF,    (6)

in which, was the dot product. Based on this, the network can adaptively attention important characteristics, noise and suppress irrelevant information. These features were then fed into a classifier to predict the final outcome.

3 Experiment and result analysis

3.1 Data sets and experimental settings

3.1.1 Dataset

This study collected 337 patients diagnosed with acute cerebral infarction and included MRI examinations from January 2019 to January 2023 in Panyu Central Hospital of Guangzhou. Please refer to Table 1 and Supplementary material for details of the data set.

TABLE 1
www.frontiersin.org

Table 1. Distribution of cases (n = 337) and distribution of training test data (n = 3,106).

3.1.2 Experimental setup

This experiment was deployed in the PyTorch deep learning framework. The server used was equipped with two NVIDIA GeForce RTX 3090 GPUs with 24G memory. In the study, the input images are resize to 224 × 224 in both training and testing. During training, the cross-entropy loss function is applied to calculate the loss values between the predicted results and ground-truth labels. Moreover, the loss of the network is minimized by the Adam optimizer with a learning rate of 0.001, where the batch size is set to 32 and training is stopped after 200 epochs.

3.1.3 Evaluation indicators

This article selected the accuracy, precision, sensitivity (i.e., recall), F1-score as a model of evaluation index, the calculation formula, respectively, as follows:

Accuracy=TP+TNTP+FP+TN+FN    (7)
Precision=TPTP+FP    (8)
Sensitivity=TPTP+FN    (9)
Specificity=TNTN+FP    (10)
F1score=2×Precision×SensitivityPrecision+Sensitivity    (11)

Here, true positives (TP) and true negatives (TN) are the number of positive samples (that is, poor prognosis) and negative samples (that refers to good prognosis) that are correctly predicted, respectively. False positive (FP) is the number of negative samples misjudged as poor prognosis, and false negative (FN) is the number of poor prognosis samples misjudged as negative. In all experiments, the overall performance of the proposed and comparative methods was evaluated by calculating the mean and standard deviation of the cross-validation measures.

3.2 Experimental results

3.2.1 Comparison of performance indicators of different models

To validate the performance of the proposed model, it was compared with advanced image classification frameworks, including Res2Net (20), ResNet50 (22), CSPNet (23), EfficientNet (24), HRNet (25) and VGG16-GLAM. VGG16-GLAM is regarded as the proposed MRI analysis model that only takes MRI as input. The performance metrics of different models were shown in Table 1. The proposed model achieved an accuracy of 91.17%, precision of 87.26%, recall of 95.21%, and F1-score of 91.06%, which significantly outperformed existing CNN frameworks. Compared to ResNet50 (22), the proposed model showed improvements of 6.04% in accuracy, 20.97% in recall, and 4.65% in F1-score. Although our method did not achieve optimal precision, it significantly improved the prediction of adverse prognosis in stroke patients compared to other methods. This has important clinical implications in the treatment of stroke patients, and it confirmed that the fusion of multidimensional data information, including clinical data, radiological features, treatment plans, prognosis, and complications, was beneficial for assessing the quality of life in patients with acute ischemic stroke.

Furthermore, the Grad-CAM technique (26) was applied to visualize the proposed model. As observed in Figure 4, the Grad-CAM saliency maps of ResNet50 can roughly locate the lesion regions, but they may overlook or misjudge certain small lesions (e.g., the last two images in the last column). In contrast, the proposed model can accurately capture lesions of various complex shapes, even detecting and paying attention to minor abnormalities. This indicated that by learning joint features that encompass global context and local refinement, the model could adaptively detect relevant lesion regions to extract more discriminative feature representations, leading to better identification of patient case types, surpassing other state-of-the-art techniques. The experimental resulted in Figure 4 strongly support that the proposed model is intuitive and interpretable, confirming that the model’s decisions primarily depend on the lesion regions while disregarding irrelevant image content.

FIGURE 4
www.frontiersin.org

Figure 4. Examples of Grad-CAM thermal map of MRI. Original images (top row) and the Grad-CAM thermal map of the ResNet50 model (2nd row) and the proposed method (bottom row).

3.2.2 Ablation experiments

In this subsection, we conducted a series of ablation experiments to validate the effectiveness of each key component of the prognostic assessment model for acute ischemic stroke based on multidimensional radiological data. Ablation studies for the proposed modules, including the multi-scale (MS) framework, GLAM, and the MLP. The No. 1 setting is the baseline network (i.e., VGG16) without these three modules, and then adding the to it as the No. 2 setting. GLAM is integrated into No. 2 to verify its effectiveness (No. 3). The No. 4 setting is the full version of the proposed network (see Table 2).

TABLE 2
www.frontiersin.org

Table 2. Comparison of the performance indicators of the different models.

3.2.2.1 Effectiveness of the multiscale framework

To explore the contribution of the multiscale framework, we first used a single-scale network as the baseline model (No. 1) and compared it with other configurations. Table 3 showed that the proposed multiscale framework could learn better feature representations, resulting in performance improvements in all metrics compared to the single-scale framework. The No. 2 configuration using the multiscale framework achieved an accuracy improvement of 5.6%, sensitivity improvement of 12.12%, and precision improvement of 2.1%. The AUC of No. 1 and No. 2 was 0.843 and 0.894 respectively, and the model quality increased from 0.82 to 0.87. As seen in Figures 5A,B,E,F,I,J, the multiscale framework significantly enhanced the prediction of adverse prognosis in stroke patients. These results indicated that the multiscale framework was capable of capturing lesions of different scales and shapes, effectively leveraging the overall characteristics of brain MRI images.

TABLE 3
www.frontiersin.org

Table 3. Effectiveness of component of the model.

FIGURE 5
www.frontiersin.org

Figure 5. Predictive model evaluation. (A–D) Receiver Operating Characteristic curve. (E–H) Overall model quality. (I–L) Confusion matrixes of the ablation experiment. P: Poor prognosis; N: Eusemia.

3.2.2.2 Effectiveness of GLAM

Compared to No. 2, the inclusion of the GLAM module in No. 3 effectively improved the recognition performance, with an accuracy improvement of 1.35%, sensitivity improvement of 0.51%, and precision improvement of 2.37%. The AUC of No. 2 and No. 3 was 0.894 and 0.908 respectively, and the model quality increased from 0.88 to 0.89. As shown in Figures 5B,C,F,G,J,K, the addition of the GLAM module enhanced the negative prediction level for stroke patients compared to Figure 5J. This indicated that the GAB in GLAM models the global contextual information, constructing global-guided attention to adaptively focus on affected regions while disregarding irrelevant information for better lesion identification. The LAB could adaptively refine local features to obtain detailed lesion information. The combination of these two modules effectively utilized joint characteristics to adaptively focus on lesion regions and learned more detailed imaging features, thereby achieving better diagnostic performance. This could serve as the foundation for clinical prognosis assessment in stroke patients.

3.2.2.3 Effectiveness of the patient metadata encoder based on MLP

Compared to No. 3, the fusion of patient metadata in No. 4 effectively improved the recognition performance, with an accuracy improvement of 0.89%, sensitivity improvement of 7.08%, and F1-score improvement of 1.22%. The AUC of No. 3 and No. 4 was 0.908and 0.916 respectively, and the model quality was 0.89. As observed in the comparison between Figures 5C,D,G,H,K,L, while there was a slight decrease in the negative prediction level, there was a significant improvement in the positive prediction level for stroke patients. In terms of clinical significance, the improved accuracy in positive prediction was more meaningful than negative prediction. This indicated that the inclusion of multidimensional clinical information could provide higher quality and more accurate assessments for patients.

4 Discussion

In the past few decades, various machine learning techniques, including logistic regression (LR) (27), linear discriminant analysis (LDA) (28), support vector machines (SVM) (29), decision trees (DT) (30), random forests (RF) (31), and neural networks (32), have been applied. These methods heavily rely on feature engineering, such as shape, texture, and pixel intensity distribution (histogram) obtained from computer programs, which can be used to identify potential imaging-based biomarkers and serve as input for improved machine learning models (33). SVM has improved the identification of carotid atherosclerosis (CA) from magnetic resonance brain images and prevented ischemic stroke in 97.5% of patients (34). The RF algorithm combined with geodesic active contour (GAC) model can automatically segment cerebrospinal fluid (CSF) in CT images for early identification of brain edema, a major medical complication after ischemic stroke (35). The LR algorithm has aided in the analysis of CT angiography (CTA) lesions and the discrimination of mobile intraluminal thrombus and atherosclerotic plaques, assisting in the selection of stroke treatment plans, with a sensitivity of 87.5% (36). The use of artificial neural networks to predict inadequate perfusion and the presence of effective collateral circulation in CT perfusion scans can facilitate further treatment, achieving an accuracy of 85.8% in testing on CT perfusion images of 396 patients (37). Several studies have employed machine learning methods on various public datasets to address various stroke-related issues for better improvement of healthcare systems and stroke treatment plans (38). However, the traditional machine learning approaches require preprocessing of input features and manual extraction. Optimization of image features and susceptibility to interference from multimodal imaging need further exploration and improvement (39). Recently, deep learning, as an emerging artificial intelligence (AI) technique, has the ability to automatically capture hierarchical and complex features from raw input data (4042). Deep neural networks, with multiple layers, simulate the perception of the human brain, transforming “low-level” to “high-level” representations, particularly in large-scale task solutions, in imaging classification, natural language processing, or bioinformatics (43, 44). In recent years, medical image processing has emerged as a hot research topic in deep learning, involving disease classification (45), lesion localization and segmentation, imaging reconstruction (46), and other tasks. Therefore, deep learning has been widely applied in stroke diagnosis and management, such as predicting clinical prognosis in AIS patients (47). Compared to traditional machine learning methods, deep CNN learning does not rely on handcrafted features. It automatically extracts and represents complex features when locating the core stroke lesions in CT or MRI (48). Deep learning not only saves time and effort but also captures pixel-level information of the lesions, contributing to improved diagnostic accuracy and prognosis (49).

Various artificial intelligence models have been widely applied in predicting the prognosis of ischemic stroke patients. Compared to predicting future stroke lesions on CT or MRI predicting patient prognosis is more challenging because commonly used prognostic scoring systems, such as the modified Rankin scale (mRS), are nonlinear and subjective, analyzing patients as a whole rather than on a voxel-by-voxel basis. This means there are fewer opportunities for artificial neural networks to learn from data, requiring larger training datasets to compensate for this limitation.

Previous studies mostly used non-imaging data as input and employed simple statistical models or machine learning models to predict prognosis (47, 5053). However, CT or MRI can provide more information such as the size and location of infarctions. Tang et al. (54) utilized machine learning techniques combined with clinical data and the core-penumbra mismatch ratio from MRI and MRI perfusion to determine post-thrombolysis clinical outcomes. The short-term (7-day) result had an area under curve (AUC) of 0.863 [95% confidence interval (CI), 0.774–0.951], and the long-term (90 days) result had an AUC of 0.778 (95% CI, 0.668–0.888). Decision tree-based algorithms were able to predict the recovery outcomes (mRS >2 at 90 days) utilizing imaging and clinical data, with AUCs of 0.746 (extreme gradient boosting) and 0.748 (gradient boosting machine). Wang et al. (9) and Zhou et al. (8) used a multivariate logistic regression model to construct an imaging omics nomogram containing patient characteristics and imaging omics characteristics, and the AUC used to predict stroke outcome was greater than 0.80. Sun et al. (10) used clinical features and apparent diffusion coefficient maps to predict poor prognosis of acute stroke (mRS score >2) and the AUC was 0.80. These models were superior to models using non-imaging data, and the clinical data were continuous and related, which demonstrates the great potential of the combination. The performance of this algorithm further improved when National Institute of Health stroke scale (NIHSS) 24 h and reperfusion status were included (55). Machine learning techniques, including regularized logistic regression, linear support vector machine, and random forest, outperformed existing pre-treatment scoring methods in predicting favorable clinical outcomes (90 days mRS >2) for patients undergoing thrombectomy for large vessel occlusion (LVO) (50).

Osama et al. (56) developed a parallel multi-parameter feature embedding Siamese network (PMFE-SN) that can learn from a small number of samples and handle skewness in multi-parameter MRI data. The proposed multi-parameter embedding architecture in PMFE-SN is based on deep learning and avoids overfitting even with a small number of samples in the dataset. The authors successfully predicted the prognosis of acute ischemic stroke patients 3 months later using MRI perfusion images and clinical data from the 2017 Ischemic Stroke LEsion Segmentation (ISLES) challenge, demonstrating superior performance compared to other advanced techniques.

Hilbert et al. (57) compared a deep learning model constructed using residual neural networks with a machine learning model utilizing traditional radiological image markers. The results showed that automatic image analysis using deep learning methods outperformed previous radiological image markers in predicting the prognosis of ischemic stroke patients and had the potential to improve treatment selection.

The proposed multi-dimensional ADL scoring network framework for AIS patients has higher accuracy than other state-of-the-art network models. Ablation experiments also confirmed the effectiveness of each module in the framework. In addition, the visualization results using the Grad-CAM technique show that our method can accurately locate the lesion area while ignoring irrelevant background information, indicating that the final identification results determined by the model are reliable and interpretable. This will help to provide more rich and comprehensive information for providing clinical diagnosis and treatment.

The effect of allowing machines to autonomously learn to fit nonlinear equations based on massive data rather than artificial formulas is closer to the complex problem itself. The same deep learning network can be trained for different types of samples and produce different fitting models individually to enhance its general applicability. Therefore, the development of acute cerebral infarction prognostic risk prediction models based on imaging and multi-dimensional data based on PyTorch deep learning framework is of great significance for early evaluation and intervention, guiding treatment plans and judging prognosis, reducing disability rate and reducing social and economic burden.

A total of 337 patients were included in our study, including 237 in the training set and 100 in the testing set. A study by Quan et al. (58), such as using fluid attenuated inversion recovery (FLAIR) and apparent diffusion coefficient (ADC) images to extract the image of omics characteristics to predict the prognosis of patients with AIS, included 190 cases of acute ischemic stroke patients, divided into the training group (n = 110) and external validation group (n = 80). In the study by Tang et al. (59) to predict the prognosis of patients with acute ischemic stroke receiving thrombolytic therapy, 168 patients with acute ischemic stroke were included. Compared with these studies on predicting the prognosis of ischemic stroke, our study not only included a large sample size, but also had a richer content, not limited to a specific treatment (thrombolysis, mechanical thrombectomy, etc.) and a single subtype of ischemic stroke. The larger sample size in our study provided more statistical power and enhances the reliability of the findings. This broader scope allowed for a more comprehensive understanding of the factors influencing the prognosis of ischemic stroke patients.

However, it is important to acknowledge the limitations of our study. Firstly, the retrospective nature of the study design introduces inherent limitations. Retrospective studies rely on existing data, which may be subject to selection bias and confounding factors. Prospective studies would provide more robust evidence and minimize potential biases. Secondly, all the samples used in our study were obtained from a single center, which might limit the generalizability of the findings. The patient population and treatment protocols at a single center may not be representative of other centers or populations. Therefore, multicenter validation is necessary to confirm the external validity and generalizability of our results.

To address these limitations, future studies could employ prospective designs with larger and more diverse samples, including patients from multiple centers. In addition, the inclusion of additional clinical and imaging variables could further enhance the predictive accuracy of our models. By addressing these limitations and conducting more rigorous studies, we can strengthen the evidence base and improve the prediction of prognosis in patients with ischemic stroke.

5 Conclusion

In this study, a model consisted of ALAM, MLP-FE, CMAM were constructed for multi-dimensional data information fusion, such as clinical data, imaging features, treatment plans, prognosis and complications, so as to achieve complementary advantages. The fusion features not only retain the MRI images features, but also make up for the clinical relevant information of the patient, which provided higher quality, more accurate and more personalized medical diagnostic for prognosis of AIS.

Data availability statememt

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding authors.

Ethics statement

The Ethics Committee of Panyu District Central Hospital of Guangzhou reviewed and approved this study. The ethics committee waived the requirement of written informed consent for participation due to the retrospective nature of the study. The study was conducted in accordance with the local legislation and institutional requirements.

Author contributions

JL: Data curation, Writing – original draft. JF: Methodology, Writing – original draft. ZL: Methodology, Writing – original draft. JW: Investigation, Writing – original draft. XL: Software, Writing – original draft. QW: Data curation, Writing – original draft. BH: Validation, Writing. HC: Project administration, Writing – original draft, Writing – review & editing. YY: Formal analysis, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported by the Basic and Applied Basic Research Foundation of Guangdong Province (No. 2022A1515220217), Medical Scientific Research Foundation of Guangdong Province (Nos. A2023216 and A2022524), Science and Technology Program of Guangzhou (Nos. 202201010840, 202201010810, 202102080532, 202002030032, and 202002020023), Health Commission Program of Guangzhou (Nos. 20212A010025 and 20201A010085), Internal scientific research fund of Guangzhou Panyu Central Hospital (Nos. 2021Z001), Panyu General Science and Technology Plan Project (Nos. 2022-Z04-006), Panyu Major Science and Technology Plan Project (Nos. 2022-Z04-114).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fneur.2023.1294723/full#supplementary-material

References

1. Strong, K, Mathers, C, and Bonita, R. Preventing stroke: saving lives around the world. Lancet Neurol. (2007) 6:182–7. doi: 10.1016/S1474-4422(07)70031-5

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Hosny, A, Parmar, C, Quackenbush, J, Schwartz, LH, and Aerts, HJWL. Artificial intelligence in radiology. Nat Rev Cancer. (2018) 18:500–10. doi: 10.1038/s41568-018-0016-5

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Panayides, AS, Amini, A, Filipovic, ND, Sharma, A, Tsaftaris, SA, Young, A, et al. AI in medical imaging informatics: current challenges and future directions. IEEE J Biomed Health Inform. (2020) 24:1837–57. doi: 10.1109/JBHI.2020.2991043

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Ge, Y, Wang, Q, Wang, L, Wu, H, Peng, C, Wang, J, et al. Predicting post-stroke pneumonia using deep neural network approaches. Int J Med Inform. (2019) 132:103986. doi: 10.1016/j.ijmedinf.2019.103986

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Li, X, Wu, M, Sun, C, Zhao, Z, Wang, F, Zheng, X, et al. Using machine learning to predict stroke-associated pneumonia in Chinese acute ischaemic stroke patients. Eur J Neurol. (2020) 27:1656–63. doi: 10.1111/ene.14295

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Flint, AC, Cullen, SP, Rao, VA, Faigeles, BS, Pereira, VM, Levy, EI, et al. The THRIVE score strongly predicts outcomes in patients treated with the solitaire device in the SWIFT and STAR trials. Int J Stroke. (2014) 9:698–704. doi: 10.1111/ijs.12292

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Papavasileiou, V, Milionis, H, Michel, P, Makaritsis, K, Vemmou, A, Koroboki, E, et al. ASTRAL score predicts 5 years dependence and mortality in acute ischemic stroke. Stroke. (2013) 44:1616–20. doi: 10.1161/STROKEAHA.113.001047

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Zhou, Y, Wu, D, Yan, S, Xie, Y, Zhang, S, Lv, W, et al. Feasibility of a clinical-radiomics model to predict the outcomes of acute ischemic stroke. Korean J Radiol. (2022) 23:811–20. doi: 10.3348/kjr.2022.0160

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Wang, H, Sun, Y, Ge, Y, Wu, PY, Lin, J, Zhao, J, et al. A clinical-radiomics nomogram for functional outcome predictions in ischemic stroke. Neurol Ther. (2021) 10:819–32. doi: 10.1007/s40120-021-00263-2

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Sun, Y, Zhuang, Y, Zhu, J, Song, B, and Wang, H. Texture analysis of apparent diffusion coefficient maps in predicting the clinical functional outcomes of acute ischemic stroke. Front Neurol. (2023) 11:1132318. doi: 10.3389/fneur.2023.1132318

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Dragoș, HM, Stan, A, Pintican, R, Feier, D, Lebovici, A, Panaitescu, PȘ, et al. MRI radiomics and predictive models in assessing ischemic stroke outcome—a systematic review. Diagnostics. (2023) 13:857. doi: 10.3390/diagnostics13050857

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Lin, Z, He, Z, Xie, S, Wang, X, Tan, J, Lu, J, et al. AANet: adaptive attention network for COVID-19 detection from chest X-ray images. IEEE Trans Neural Netw Learn Syst. (2021) 32:4781–92. doi: 10.1109/TNNLS.2021.3114747

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Shi, M, Shen, J, Yi, Q, Weng, J, Huang, Z, Luo, A, et al. LMFFNet: a well-balanced lightweight network for fast and accurate semantic segmentation. IEEE Trans Neural Netw Learn Syst. (2023) 34:3205–19. doi: 10.1109/TNNLS.2022.3176493

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Vaswani, A, Shazeer, N, Parmar, N, Uszkoreit, J, Jones, L, Gomez, AN, et al. Attention is all you need, Proceedings of the 31st International Conference on Neural Information Processing Systems (2017) 6000–6010

Google Scholar

15. Dosovitskiy, A, Beyer, L, Kolesnikov, A, Weissenborn, D, Zhai, X, Unterthiner, T, et al. An image is worth 16 × 16 words: transformers for image recognition at scale. International Conference on Learning Representations (2020)

Google Scholar

16. Lin, Z, He, Z, et al. DBGANet: dual-branch geometric attention network for accurate 3D tooth segmentation. IEEE Trans Circuits Syst Video Technol. (2023):1. doi: 10.1109/TCSVT.2023.3331589

CrossRef Full Text | Google Scholar

17. Mu, N, Wang, H, Zhang, Y, Jiang, J, and Tang, J. Progressive global perception and local polishing network for lung infection segmentation of COVID-19 CT images. Pattern Recogn. (2021) 120:108168. doi: 10.1016/j.patcog.2021.108168

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Lin, Z, He, Z, Yao, R, Wang, X, Liu, T, Deng, Y, et al. Deep dual attention network for precise diagnosis of COVID-19 from chest ct images. IEEE Trans Artif Intell. (2022) 1:1–11. doi: 10.1109/TAI.2022.3225372

CrossRef Full Text | Google Scholar

19. Yan, Q, Wang, B, Zhang, W, Luo, C, Xu, W, Xu, Z, et al. Attention-guided deep neural network with multi-scale feature fusion for liver vessel segmentation. IEEE J Biomed Health Inform. (2021) 25:2629–42. doi: 10.1109/JBHI.2020.3042069

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Gao, SH, Cheng, MM, Zhao, K, Zhang, XY, Yang, MH, and Torr, P. Res2Net: a new multi-scale backbone architecture. IEEE Trans Pattern Anal Mach Intell. (2021) 43:652–62. doi: 10.1109/TPAMI.2019.2938758

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Simonyan, K, and Zisserman, A. Very deep convolutional networks for large-scale image recognition Proceedings of 3rd International Conference on Learning Representations (ICLR) (2015), pp. 1–14

Google Scholar

22. He, K, Zhang, X, Ren, S, and Sun, J. Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, (2016), pp. 770–778

Google Scholar

23. Wang, CY, Mark Liao, HY, Wu, YH, Chen, PY, Hsieh, JW, and Yeh, IH. CSPNet: a new backbone that can enhance learning capability of CNN. Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2020. IEEE Computer Society. (2020)1571–1580

Google Scholar

24. Tan, M, and Le, Q. EfficientNet: rethinking model scaling for convolutional neural network. Proceedings of International Conference on Machine Learning (ICML). (2019): 6105–6114

Google Scholar

25. Wang, J, Sun, K, Cheng, T, Jiang, B, Deng, C, Zhao, Y, et al. Deep high-resolution representation learning for visual recognition. IEEE Trans Pattern Anal Mach Intell. (2021) 43:3349–64. doi: 10.1109/TPAMI.2020.2983686

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Selvaraju, RR, Cogswell, M, Das, A, Vedantam, R, Parikh, D, and Batra, D Grad-CAM: visual explanations from deep networks via gradient-based localization. Proceedings of the IEEE International Conference on Computer Vision. (2017): 618–626

Google Scholar

27. Kleinbaum, DG, Dietz, K, Gail, M, and Klein, M. Logistic regression. New York, NY: Springer-Verlag (2002).

Google Scholar

28. Xanthopoulos, P, Pardalos, PM, and Trafalis, TB. Robust data mining. London: Springer (2013).

Google Scholar

29. Hearst, MA, Dumais, ST, Osuna, E, Platt, J, and Scholkopf, B. Support vector machines. IEEE Intell Syst Appl. (1998) 13:18–28. doi: 10.1109/5254.708428

CrossRef Full Text | Google Scholar

30. Rokach, L, and Maimon, O. Decision trees In: Data mining and knowledge discovery handbook. Boston, MA: Springer (2005). 165–92.

Google Scholar

31. Breiman, L. Random forests. Mach Learn. (2001) 45:5–32. doi: 10.1023/A:1010933404324

CrossRef Full Text | Google Scholar

32. Bishop, CM. Neural networks and their applications. Rev Sci Instrum. (1994) 65:1803–32. doi: 10.1063/1.1144830

CrossRef Full Text | Google Scholar

33. Castaneda-Vega, S, Katiyar, P, Russo, F, et al. Machine learning identifies stroke features between species. Theranostics. (2021) 11:3017–34. doi: 10.7150/thno.51887

PubMed Abstract | CrossRef Full Text | Google Scholar

34. Bento, M, Souza, R, Salluzzi, M, Rittner, L, Zhang, Y, and Frayne, R. Automatic identification of atherosclerosis subjects in a heterogeneous MR brain imaging data set. Magn Reson Imaging. (2019) 62:18–27. doi: 10.1016/j.mri.2019.06.007

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Chen, Y, Dhar, R, Heitsch, L, Ford, A, Fernandez-Cadenas, I, Carrera, C, et al. Automated quantification of cerebral edema following hemispheric infarction: application of a machine-learning algorithm to evaluate CSF shifts on serial head CTs. Neuroimage Clin. (2016) 12:673–80. doi: 10.1016/j.nicl.2016.09.018

PubMed Abstract | CrossRef Full Text | Google Scholar

36. Thornhill, RE, Lum, C, Jaberi, A, Stefanski, P, Torres, CH, Momoli, F, et al. Can shape analysis differentiate free-floating internal carotid artery thrombus from atherosclerotic plaque in patients evaluated with CTA for stroke or transient ischemic attack? Acad Radiol. (2014) 21:345–54. doi: 10.1016/j.acra.2013.11.011

CrossRef Full Text | Google Scholar

37. Vargas, J, Spiotta, A, and Chatterjee, AR. Initial experiences with artificial neural networks in the detection of computed tomography perfusion deficits. World Neurosurg. (2018) 124:e10–6. doi: 10.1016/j.wneu.2018.10.084

PubMed Abstract | CrossRef Full Text | Google Scholar

38. Sirsat, MS, Fermé, E, and Câmara, J. Machine learning for brain stroke: a review. J Stroke Cerebrovasc Dis. (2020) 29:105162. doi: 10.1016/j.jstrokecerebrovasdis.2020.105162

CrossRef Full Text | Google Scholar

39. Sheth, SA, Giancardo, L, Colasurdo, M, Srinivasan, VM, Niktabe, A, and Kan, P. Machine learning and acute stroke imaging. J Neurointerv Surg. (2023) 15:195–9. doi: 10.1136/neurintsurg-2021-018142

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Chan, HP, Samala, RK, Hadjiiski, LM, and Zhou, C. Deep learning in medical image analysis. Adv Exp Med Biol. (2020) 1213:3–21. doi: 10.1007/978-3-030-33128-3_1

CrossRef Full Text | Google Scholar

41. Puttagunta, M, and Ravi, S. Medical image analysis based on deep learning approach. Multimed Tools Appl. (2021) 80:24365–98. doi: 10.1007/s11042-021-10707-4

PubMed Abstract | CrossRef Full Text | Google Scholar

42. Alam, M, Samad, MD, Vidyaratne, L, Glandon, A, and Iftekharuddin, KM. Survey on deep neural networks in speech and vision systems. Neurocomputing. (2020) 417:302–21. doi: 10.1016/j.neucom.2020.07.053

PubMed Abstract | CrossRef Full Text | Google Scholar

43. Akkus, Z, Cai, J, Boonrod, A, Zeinoddini, A, Weston, AD, Philbrick, KA, et al. A survey of deep-learning applications in ultrasound: artificial intelligence-powered ultrasound for improving clinical workflow. J Am Coll Radiol. (2019) 16:1318–28. doi: 10.1016/j.jacr.2019.06.004

PubMed Abstract | CrossRef Full Text | Google Scholar

44. LeCun, Y, Bengio, Y, and Hinton, G. Deep learning. Nature. (2015) 521:436–44. doi: 10.1038/nature14539

PubMed Abstract | CrossRef Full Text | Google Scholar

45. Zhang, J, Xie, Y, Wu, Q, and Xia, Y. Medical image classification using synergic deep learning. Med Image Anal. (2019) 54:10–9. doi: 10.1016/j.media.2019.02.010

PubMed Abstract | CrossRef Full Text | Google Scholar

46. Kim, M, Yun, J, Cho, Y, et al. Deep learning in medical imaging. Neurospine. (2020) 17:471–2. doi: 10.14245/ns.1938396.198.c1

PubMed Abstract | CrossRef Full Text | Google Scholar

47. van Os, HJA, Ramos, LA, Hilbert, A, van Leeuwen, M, van Walderveen, MAA, Kruyt, ND, et al. Predicting outcome of endovascular treatment for acute ischemic stroke: potential value of machine learning algorithms. Front Neurol. (2018) 9:784. doi: 10.3389/fneur.2018.00784

PubMed Abstract | CrossRef Full Text | Google Scholar

48. Li, L, Wei, M, Liu, B, Atchaneeyasakul, K, Zhou, F, Pan, Z, et al. Deep learning for hemorrhagic lesion detection and segmentation on brain CT images. IEEE J Biomed Health Inform. (2021) 25:1646–59. doi: 10.1109/JBHI.2020.3028243

PubMed Abstract | CrossRef Full Text | Google Scholar

49. Zhang, S, Zhang, M, Ma, S, Wang, Q, Qu, Y, Sun, Z, et al. Research progress of deep learning in the diagnosis and prevention of stroke. Biomed Res Int. (2021) 2021:5213550–5. doi: 10.1155/2021/5213550

PubMed Abstract | CrossRef Full Text | Google Scholar

50. Nishi, H, Oishi, N, Ishii, A, Ono, I, Ogura, T, Sunohara, T, et al. Predicting clinical outcomes of large vessel occlusion before mechanical thrombectomy using machine learning. Stroke. (2019) 50:2379–88. doi: 10.1161/STROKEAHA.119.025411

PubMed Abstract | CrossRef Full Text | Google Scholar

51. Heo, J, Yoon, JG, Park, H, Kim, YD, Nam, HS, and Heo, JH. Machine learning-based model for prediction of outcomes in acute stroke. Stroke. (2019) 50:1263–5. doi: 10.1161/STROKEAHA.118.024293

CrossRef Full Text | Google Scholar

52. Ntaios, G, Faouzi, M, Ferrari, J, Lang, W, Vemmos, K, and Michel, P. An integer-based score to predict functional outcome in acute ischemic stroke: the ASTRAL score. Neurology. (2012) 78:1916–22. doi: 10.1212/WNL.0b013e318259e221

PubMed Abstract | CrossRef Full Text | Google Scholar

53. Ho, KC, Speier, W, El-Saden, S, Liebeskind, DS, Saver, JL, Bui, AA, et al. Predicting discharge mortality after acute ischemic stroke using balanced data. AMIA Annu Symp Proc. (2014) 2014:1787–96.

Google Scholar

54. Tang, TY, Jiao, Y, Cui, Y, Zeng, CH, Zhao, DL, Zhang, Y, et al. Development and validation of a penumbra-based predictive model for thrombolysis outcome in acute ischemic stroke patients. EBioMedicine. (2018) 35:251–9. doi: 10.1016/j.ebiom.2018.07.028

PubMed Abstract | CrossRef Full Text | Google Scholar

55. Xie, Y, Jiang, B, Gong, E, Li, Y, Zhu, G, Michel, P, et al. JOURNAL CLUB: use of gradient boosting machine learning to predict patient outcome in acute ischemic stroke on the basis of imaging, demographic, and clinical information. AJR Am J Roentgenol. (2019) 212:44–51. doi: 10.2214/AJR.18.20260

PubMed Abstract | CrossRef Full Text | Google Scholar

56. Osama, S, Zafar, K, and Sadiq, MU. Predicting clinical outcome in acute ischemic stroke using parallel multi-parametric feature embedded Siamese network. Diagnostics. (2020) 10:858. doi: 10.3390/diagnostics10110858

PubMed Abstract | CrossRef Full Text | Google Scholar

57. Hilbert, A, Ramos, LA, van Os, HJA, Olabarriaga, SD, Tolhuisen, ML, Wermer, MJH, et al. Data-efficient deep learning of radiological image data for outcome prediction after endovascular treatment of patients with acute ischemic stroke. Comput Biol Med. (2019) 115:103516. doi: 10.1016/j.compbiomed.2019.103516

PubMed Abstract | CrossRef Full Text | Google Scholar

58. Quan, G, Ban, R, Ren, JL, Liu, Y, Wang, W, Dai, S, et al. FLAIR and ADC image-based radiomics features as predictive biomarkers of unfavorable outcome in patients with acute ischemic stroke. Front Neurosci. (2021) 15:730879. doi: 10.3389/fnins.2021.730879

PubMed Abstract | CrossRef Full Text | Google Scholar

59. Tang, TY, Jiao, Y, Cui, Y, Zhao, DL, Zhang, Y, Wang, Z, et al. Penumbra-based radiomics signature as prognostic biomarkers for thrombolysis of acute ischemic stroke patients: a multicenter cohort study. J Neurol. (2020) 267:1454–63. doi: 10.1007/s00415-020-09713-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: prognostic risk, assessment model, acute ischemic stroke, imaging, multidimensional data, MRI

Citation: Liang J, Feng J, Lin Z, Wei J, Luo X, Wang QM, He B, Chen H and Ye Y (2023) Research on prognostic risk assessment model for acute ischemic stroke based on imaging and multidimensional data. Front. Neurol. 14:1294723. doi: 10.3389/fneur.2023.1294723

Received: 18 September 2023; Accepted: 30 November 2023;
Published: 19 December 2023.

Edited by:

Tarun Singh, University of Michigan, United States

Reviewed by:

Ziga Spiclin, University of Ljubljana, Slovenia
Zhe Wang, The First Affiliated Hospital of Guangdong Pharmaceutical University, China
Cheng Long, South China Normal University, China

Copyright © 2023 Liang, Feng, Lin, Wei, Luo, Wang, He, Chen and Ye. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Hanwei Chen, docterwei@sina.com; chenhanwei0756@pyhospital.com.cn; Yufeng Ye, 838554325@qq.com

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.