Skip to main content

ORIGINAL RESEARCH article

Front. Neurol., 21 February 2023
Sec. Stroke
This article is part of the Research Topic Machine learning in data analysis for stroke/endovascular therapy View all 11 articles

A deep learning approach to predict collateral flow in stroke patients using radiomic features from perfusion images

  • 1Department of Computer Science, Technische Universität München, München, Germany
  • 2Neuroradiology, Klinikum Rechts der Isar, Technische Universität München, München, Germany
  • 3Institute for Diagnostic and Interventional Neuroradiology, Inselspital University Hospital, Bern, Switzerland
  • 4Department of Quantitative Biomedicine, University of Zurich, Zurich, Switzerland

Collateral circulation results from specialized anastomotic channels which are capable of providing oxygenated blood to regions with compromised blood flow caused by arterial obstruction. The quality of collateral circulation has been established as a key factor in determining the likelihood of a favorable clinical outcome and goes a long way to determining the choice of a stroke care model. Though many imaging and grading methods exist for quantifying collateral blood flow, the actual grading is mostly done through manual inspection. This approach is associated with a number of challenges. First, it is time-consuming. Second, there is a high tendency for bias and inconsistency in the final grade assigned to a patient depending on the experience level of the clinician. We present a multi-stage deep learning approach to predict collateral flow grading in stroke patients based on radiomic features extracted from MR perfusion data. First, we formulate a region of interest detection task as a reinforcement learning problem and train a deep learning network to automatically detect the occluded region within the 3D MR perfusion volumes. Second, we extract radiomic features from the obtained region of interest through local image descriptors and denoising auto-encoders. Finally, we apply a convolutional neural network and other machine learning classifiers to the extracted radiomic features to automatically predict the collateral flow grading of the given patient volume as one of three severity classes - no flow (0), moderate flow (1), and good flow (2). Results from our experiments show an overall accuracy of 72% in the three-class prediction task. With an inter-observer agreement of 16% and a maximum intra-observer agreement of 74% in a similar experiment, our automated deep learning approach demonstrates a performance comparable to expert grading, is faster than visual inspection, and eliminates the problem of grading bias.

1. Introduction

Collateral circulation results from specialized anastomotic channels which are present in most tissues and capable of providing nutrient perfusion to regions with compromised blood flow due to ischemic injuries caused by ischemic stroke, coronary atherosclerosis, peripheral artery disease, and similar conditions or diseases (1). Collateral circulation helps to sustain blood flow in the ischaemic areas in acute, subacute, or chronic phases after an ischaemic stroke or transient ischaemic attack (2). The quality of collateral circulation has been convincingly established as a key factor in determining the likelihood of successful reperfusion and favorable clinical outcome (3). It is also seen as one of the major determinants of infarct growth in the early time windows which is likely to have an impact on the chosen stroke care model that is the decision to transport or treat eligible patients immediately.

A high number of imaging methods exist to assess the structure of the cerebral collateral circulation and several grading criteria have been proposed to quantify the characteristics of collateral blood flow. However, this grading is mostly done through visual inspection of the acquired images which introduces two main challenges.

First, there are biases and inconsistencies in the current grading approaches: There is a high tendency of introducing bias in the final grade assigned to a patient depending on the experience level of the clinician. There are inconsistencies also in the grade assigned by a particular clinician at different times for the same patient. These inconsistencies are quantified at 16% interobserver agreement and a maximum intraobserver agreement of 74% respectively in a similar study by Ben Hassen et al. (4).

Second, grading is time-consuming and tedious: Aside the problem of bias prediction, it also takes the clinician several minutes to go through the patient images to first select the correct image sequence, detect the region of collateral flow and then to be able to assign a grading a period of time which otherwise could have been invested in the treatment of the patient.

In this work, we analyze several machine learning and deep learning strategies that aim toward automating the process of collateral circulation grading. We present a set of solutions focusing on two main aspects of the task at hand.

First, the region of interest needs to be identified. We automate the extraction of the region of interest (ROI) from the patient images using deep reinforcement learning (RL). This is necessary for achieving a fully automated system that will require no human interaction and save the clinician the time spent on performing this task.

Finally, the region of interest needs to be processed and classified. We consider various feature extraction schemes and classifiers suitable for the task described above. This helps to extract useful image features, both learned and hand-crafted, which are relevant to the classification task. We predict digitally subtracted angiography (DSA) based collateral flow grading from MR perfusion images in this task. This saves the time required in choosing the right DSA sequence from the multiple DSA sequences acquired and helps achieve a fully automated system.

1.1. Prior work and open challenges

1.1.1. Imaging criteria for cerebral collateral circulation

Imaging methods for assessing cerebral collateral flow can be grouped under two main classification schemes, invasive vs. non-invasive and structural vs. functional imaging. Structural imaging methods provide information about the underlying structure of the cerebral collateral circulation network. Some of the commonly used structural imaging modalities are traditional single-phase computed tomography angiography (CTA), time-of-flight magnetic resonance angiography (TOF-MRA), and digitally subtracted angiography (DSA), among others. Other imaging modalities have been used in clinical practice and relevant research areas in accessing the structure of the cerebral collateral circulation are discussed in Liu et al. (2), McVerry et al. (5), Martinon et al. (6). DSA is the gold standard for assessing the collateral flow, however, due to the associated high cost and invasive nature, other non-invasive methods like CTA and MRA are commonly used (2).

Functional imaging methods are used to assess the function of the underlying cerebral collateral circulation. Single-photon emission CT (SPECT), MR perfusion, and positron emission tomography (PET) are examples of imaging methods that provide functional information about the cerebral collateral flow. MR perfusion imaging is often followed by a post-processing step to extract parametric information. Very common parametric information includes the time-to-peak (Tmax) time taken for the blood flow to reach its peak (max) at a given region in the brain, relative blood flow (rBF) volume of blood flowing through a given brain tissue per unit of time, and relative blood volume (rBV) volume of blood in a given brain tissue relative to an internal control (e.g. normal white matter or an arterial input function). Functional imaging is sometimes combined with structural imaging either in a single scanning procedure or separate procedures and can serve as complementing information in the decision making process. Here, structural imaging is oftentimes used to map the anatomy and probe tissue microstructure.

MRI perfusion and diffusion have evolved as key biomarkers in determining collateralization of stroke patients, and a patient stratification based on these markers has been proposed repeatedly (7). At the same time, a qualitative CTA and DSA based grading are the most common approaches for evaluating collateralization (810).

1.1.2. Cerebral collateral flow grading

Cerebral collateral circulation plays an important role in stabilizing cerebral blood flow when the normal blood circulation system is compromised in cases of acute, subacute, or chronic ischaemic stroke. The quality of the cerebral collateral circulation system is one of the factors that determine the speed of infarct growth and the outcome of stroke treatment and reperfusion therapies. A poor collateral flow is associated with worse outcomes and faster growth of infarcts while a good collateral flow is associated with good outcomes and slower growth of infarcts in acute stroke treatment (11). Due to the important role played by cerebral collateral blood flow, various grading scales and their association with risk factors and treatment outcomes have been discussed extensively in literature.

Several grading systems have been proposed for assessing the quality of the collateral circulation network. Among these grading systems, the DSA based system proposed by the American Society of Interventional and Therapeutic Neuroradiology/Society of Interventional Radiology (ASITN/SIR) is the most widely accepted scheme. This grading system describes the collateral flow as one of five levels of flow which are; absence of collaterals (0), slow collaterals (1), rapid collaterals (2), partial collaterals (3), and complete collaterals (4) to the periphery of the ischaemic site (2, 12). In most studies that use the ASITN/SIR scheme, the grading scale is merged into three levels—grades 0–1 (poor), 2 (moderate) and 3–4 (good collateral) flow. CTA based systems also have several grading schemes ranging from two (good, bad) to five (absent, diminished >50%, <50%, equal, more) labels (12).

The relationship between pretreatment collateral grade and vascular recanalization has been assessed for patients who received endovascular therapy for acute cerebral ischemia from two distinct study populations by Bang et al. (13). The study showed that 14.1, 25.2, and 41.5% of patients with poor, good, and excellent pretreatment collaterals respectively achieved complete revascularization. Another study by Bang et al. (14) on the relationship between MRI diffusion and perfusion lesion indices, angiographic collateral grade, and infarct growth showed that the greatest infarct growth occurred in patients with both non-recanalization and poor collaterals. Mansour (15) assessed collateral pathways in acute ischemic stroke using a new grading scale (Mansour Scale) and correlated the findings with different risk factors, clinical outcomes, and recanalization rates with endovascular management. More research (1317) has been conducted into the relationship between the cerebral collateral circulation, its grading, and the clinical outcome of the choice of treatment of acute ischemic stroke, and they all confirm a positive association between collateral flow and the success of the outcome.

Due to the crucial role played by collateral circulation, it is a common practice in most clinical procedures to determine the quality of a patient's collateral as first-hand information toward the choice of the treatment or care model. This grading is done manually by inspecting patient scans which is time-consuming and also introduces some level of bias in the final grade assigned to a patient. Ben Hassen et al. (4) evaluated the inter-and intraobserver agreement in angiographic leptomeningeal collateral flow assessment on the ASITN/SIR scale and found an overall interobserver agreement κ = 0.16 ± 6.5 × 10−3 among 19 observers with grades 0 and 1 being associated with the best results of κ = 0.52 ± 0.001 and κ = 0.43 ± 0.004 respectively. By merging the scales into two classes, poor collaterals (grade 0, 1, or 2), versus good collaterals (grade 3 or 4), the interobserver agreement increased to κ = 0.27 ± 0.014. The same study recorded maximum intraobserver agreements of κ = 0.74 ± 0.1 and κ = 0.79 ± 0.11 for the ASITN/SIR and dichotomized scales respectively. McHugh (18) presented a study on interrater reliability and the kappa statistic as a measure of agreement and recommended a moderate interobserver agreement of 0.60 ≤ κ ≤ 0.79 as a minimum requirement for medical data and study. These results are evidence of the need to automate the collateral grading process to achieve speed and consistency in the assigned grading.

Methods for automating the grading of collateral flow have not yet been properly explored in literature. Kersten-Oertel et al. (19) presented an automated technique to compute a collateral circulation score based on differences seen in mean intensities between left and right cerebral hemispheres in 4D angiography images and found a good correlation between the computed score and radiologist score (r2 = 0.71) and good separation between good and intermediate/poor groups. Grunwald et al. (20) used a machine learning approach to categorize the degree of collateral flow in 98 patients who were eligible for mechanical thrombectomy and generated an e-CTA collateral score (CTA-CS) for each patient. The experiments showed that the e-CTA generated improved the intraclass correlation coefficient between three experienced neuroradiologists from 0.58 (0.46–0.67) to 0.77 (0.66–0.85, p = 0.003).

1.1.3. Reinforcement learning for medical imaging

Defining the region of interest (ROI) is often the first step in most image-based radiomics pipelines. This is the case because full patient scans often include artifacts and other information which are irrelevant and can affect the final outcome of the study. Therefore, most pipelines propose a manual localization of a ROI as a preprocessing step. However, it is crucial to define the ROI in an automated and reproducible fashion in other to achieve a fully automated pipeline. We propose a reinforcement learning approach for the localization of the region of interest due to increased speed and lower training data requirements compared to other supervised learning approaches.

Reinforcement learning (RL) has become one of the most active research areas in machine learning and involves the training of a machine learning agent to make a sequence of reward-based decisions toward the achievement of a goal through interaction with the environment. The idea of RL has been long applied in the field of robotics for robot vision and navigation (2123) before the topic became very popular in the image processing society. RL has been used in the general field of computer vision mainly for object detection (2426), image segmentation (27, 28), and image enhancement (2931). However, in medical imaging RL is still in the research phase with very high potential. Netto et al. (32) presented an overview of medical imaging applications applying reinforcement learning with a detailed illustration of a use case involving lung nodules classification which showed promising results. Sahba et al. (27) implemented an RL based thresholding for segmenting prostate in ultrasound images with results that showed high potential for applying RL in medical image segmentation. Alansary et al. (33) evaluated reinforcement learning agents for anatomical landmark detection by comparing fixed and multi-scale search strategies with hierarchical action steps in a coarse-to-fine manner and achieved a performance better than state-of-the-art supervised learning methods.

1.2. Main contributions

In this study, we employ parametric information (Tmax, rBF, rBV) from MR perfusion images of patients with acute ischaemic stroke and predict the three-level DSA based grading of these patients based on this functional information. We hypothesize that the rich information on blood flow visible from MRI perfusion can be used to predict collateral flow in a similar manner to DSA. Moreover, we argue that this approach, using 3D information, may even offer a more reliable biomarker than the interpretation of DSA images. As collateralization patterns are often unstable and may undergo significant changes in the course of minutes, a second estimate of the activation of collateral flow using MRI—in addition to the subsequent DSA—will offer better diagnostic information.

We explore machine learning and deep learning methods in collateral flow grading. We apply deep reinforcement learning, a variant of RL which combines the power of deep learning and reinforcement learning, to detect a rigid-sized cube around the occluded region in an acute ischemic stroke patient scan as an initial step toward the prediction of cerebral collateral flow grading. This step is necessary to automate the detection of the occluded region which improves the accuracy of the prediction. Reducing the time spent on this task and ensuring that the proposed methodology is fully automated.

We provide experiments on different feature extraction strategies including denoising autoencoder (DAE), histogram of oriented gradient (HOG), and local binary pattern (LBP). The extracted features are further utilized in a random forest (RF), K-nearest neighbor (KNN), support vector machine (SVM), and convolutional neural network (CNN) classifiers for the prediction of the collateral flow grading. We provide detailed experimental setup and results which will serve as a guide for further research in this direction.

2. Methodology

In this section, we will discuss the details of the steps we employed in predicting the collateral flow grading from MR perfusion data. Figure 1 shows an overview of the main steps involved in the classification process. The first step is the detection of the region of interest (ROI) using reinforcement learning. This step helps to narrow down the classification task to only the area which has been occluded from normal blood flow. The second step deals with extracting features from the ROI. Finally, we feed the extracted features to a set of classifiers to obtain the collateral flow grading for the given patient data.

FIGURE 1
www.frontiersin.org

Figure 1. An overview of the steps involved in predicting collateral flow grading from MR perfusion parametric data. The first step involves a region of interest detection using reinforcement learning, followed by histogram of gradient (HOG), local binary pattern (LBP), and denoising autoencoder (DEA) feature extraction schemes and then the classification step which uses random forest (RF), K-nearest neighbor (KNN), support vector machine (SVM), and convolutional neural network (CNN) classifiers.

2.1. Deep reinforcement learning for region of interest detection

The idea of reinforcement learning includes an artificial agent which is trained to interact with an environment through a sequence of reward-based decisions toward a specific goal. At every time step t, the agent takes into account its current state s and performs an action a in a set of actions A and receives a reward r which is a measure of how good or bad the action a is toward the achievement of the set goal. The aim of the agent, which is to find an optimal policy (set of states, actions, and rewards) that maximizes the future reward, can be formulated as a Markov Decision Process. Since Markov Decision Process involves a large number of possible decision points which are normally not fully observable, RL approximates the optimal decision function by iteratively sampling from the set of policies through a process known as Q-learning.

2.1.1. Q-learning

At time point t and state s, let π=aii=tt+T be a policy that is a sequence of actions needed by the agent to move from the current state s to the target. Let Qt be a future discounted reward function such that

Qt(s,π)=i=tt+Tγi-trπi,    (1)

where rπi is the reward associated with the action ai of policy π at time t = i, γ∈[0, 1] is the future reward discounting factor, and T is the number of steps needed to reach the target by the chosen policy π. At any time step t the optimal policy π* is the policy that maximizes the expected value of Qt. This can be represented by an action-value function Qt*(s) defined by

π*=Qt*(s)=maxπ𝔼[Qt(s,π)]    (2)

The optimal value function Qt*(s) obeys the Bellman equation, stating that if the optimal value Qt+1*(s) of the next state is known for all possible policies π, then the optimal behavior is to select the policy π* that maximizes the expected value of rπt+Qt+1(s, π) [which follows from setting i = t in Equation (1)]. The action-value function can therefore be estimated recursively as

Qt*(s)=maxπ𝔼[rπt+Qt+1(s,π)]    (3)

If the problem space is small enough then the set of policies and state can be fully observed and Equation (3) can be used to determine the optimal policy toward the target. However, in most cases, the problem space is too complex to explore, and hence evaluating the future reward for all possible policies is not feasible. Qt*(s) is therefore approximated by a non-linear deep network Q*(s, θ) with a set of parameters θ resulting in what is known as deep Q-learning (34).

2.1.2. Agent state, action definition, and reward function

Given a 3-D scan as the agent's environment, a state s is represented by (sx, sy, sz) which is the top-left corner of a (64 × 64 × 64) cube contained in the 3-D scan. We adopt an agent history approach which involves feeding the last four states visited by the agent to the network to prevent the agent from getting stuck in a loop. Since we have a fixed-sized cube as a state our agent's set of six actions {mu, md, ml, mr, mf, mb} is made up of only movements up, down, left, right, forward, and backward respectively which enables the agent to visit all possible locations within the volume. The agent's reward for taking an action a is a function of the intersection over union (IoU) of the target state s* and the state before (sab), and after (saa) taking the action. This is given by

Ra(saa,sab)=sign[IoU(saa,s*)-IoU(sab,s*)]    (4)

where sign is the sign function that returns −1 for all values less than 0 and 1 otherwise. This leads to a binary reward (r∈{−1, 1}) scheme which represents good and bad decisions respectively. During the training stage, the agent search sequence is terminated when the IoU of the current agent's state and the target state is greater than or equal to a predefined threshold τ. At test time the agent is terminated when a sequence of decisions leads to an oscillation [as proposed by Alansary et al. (33)], that is when the agent visits one state back and forth for a period of time.

Experiments by Alansary et al. (33) and Navarro et al. (35) show that deep reinforcement learning has superior performance in object detection as compared to classical supervised learning, especially in images with a noisy background. RL agents also require lesser training data as compared to other supervised learning methods like CNN. These proven advantages make deep reinforcement learning the right choice for our limited and noisy data.

2.2. Feature extraction and classification

Feature extraction methods are used in many machine learning tasks to either reduce the dimension of the problem or to extract information from the raw input which would otherwise not be easily extracted by the underlying classifier. In this work, we extract two main classes of features—learned features through a denoising auto-encoder (DAE), and local image descriptors made up of histogram of oriented gradients (HOG) and local binary pattern (LBP).

2.2.1. Denoising auto-encoder

An auto-encoder is an unsupervised deep learning method used for dimension reduction, feature extraction, image reconstruction or denoising and is sometimes also used as a pre-training strategy in supervised learning networks. An auto-encoder is made up of two parts: an encoder Φ:XF which maps an image xX to fxF in the features domain and a decoder Ψ:FX which maps a feature set fF to xfX. The full auto-encoder is therefore a composite function of the form ΨΦ:XX. Let y^=Ψ(Φ(x)) for a given input image xX, then the learning process of auto-encoder involves finding a pair of {Φ, Ψ} such that y^i=xi for all xiX. The encoder Φ then becomes the feature extractor which is used for extracting the needed features.

If the function Φ is invertible, then the learning process can lead to a trivial solution by just choosing Ψ to be the inverse of Φ, and Ψ°Φ becomes an identity function leading to what is known as identity-function risk. To prevent this, the input image x is first corrupted by adding noise before feeding it to Φ leading to a denoising auto-encoder. We therefore have

y^=Ψ(Φ(x~)),   x~=Υ(x)    (5)

where Υ is the random image corruption function. We approximate the encoder and decoder by deep CNNs E(x, θe) and D(f, θd) parameterized by θe and θd; respectively. Training is done through back-propagating the Mean Squared Error (MSE) of the original image x and the reconstructed image y^ given by

L=1Ni=1N(y^i-xi)2    (6)

where N is the number of images in the training set or training batch. We adopt the V-Net architecture proposed by Milletari et al. (36) and simplify it by removing the fine-grained feature forwarding, and reducing the depth of the network due to limitations on the amount of training data available. The downsampling layers of the VNET architecture represent the encoding part [E(x, θe)] of the DAE and the upsampling layers represent the decoding part [D(f, θd)] of the DAE. Figure 2 shows an overview of the simplified architecture used for extracting the DAE features.

FIGURE 2
www.frontiersin.org

Figure 2. The network architecture used for extracting the DAE features. The downsampling layer is a convolution with a stride of (2 × 2 × 2) which downsamples the input volume to half of the size on every axes. The upsampling layer is a transposed convolution with a stride of (2 × 2 × 2) which doubles the size of the input on every axes.

2.2.2. Local image descriptors and classifiers

We consider two types of local image descriptors - histograms of oriented gradients (HOG) and a local binary pattern (LBP). Given a volume X, we extract the LBP encoding of each voxel by thresholding its 3 × 3 × 3 neighborhood by the intensity value p* of the center voxel which results in 26 long bits b0, b1, b2,..., b25 where bi={1,if pip*,0 otherwise} and pi is the intensity value of the ith neighbor. We then concatenate the binary encoding to a single binary number b0b1b2...b25 and then into a decimal value which results in 225 possible binary codes. Details of the implementation until this point can be found in Heikklä and Pietikäinen (37). We group the codes into two main classes—uniform codes which have at most two binary transitions and non-uniform codes which have more than two binary transitions. A binary transition is a switch from 0 to 1 or vice versa. For example the codes 0000, 000111, 011100, and 110110 have zero, one, two, and three transitions respectively. To handle noisy data and to reduce the feature space, we group all the non-uniform codes into one class and add it to the uniform codes resulting in 927 codes instead of 225 . Finally, the histogram distribution of the individual codes is extracted as the LBP feature representation for the volume X.

We also explore the HOG feature extractor based on the method proposed in Klas¨er et al. (38). Given a volume X, we quantize gradient orientations over an icosahedron and merge opposite directions in one bin resulting in 10 gradient orientations. The gradient for each voxel xiX is obtained by convolving the 5 × 5 × 5 neighborhood of the voxel by gradient filters kx, ky, and kz of the same size, giving us a gradient vector xi3. The gradient filters are zero everywhere except for the center columns along the respective axes kx(i, 3, 3) = ky(3, i, 3) = kz(3, 3, i) = [1, 0, −2, 0, 1] for i∈{1, 2,..., 5}. The gradient vectors xi are then projected to the gradient orientations and a histogram representation of these orientations are obtained and used as the HOG feature representation of the volume X.

We run experiments with four machine learning classifiers on each of the features extracted. We implement Convolutional Neural Network (CNN), Random Forest (RF), Support Vector Machine (SVM), and K-Nearest Neighbor (KNN) classifiers. Our CNN classifier in Figure 3 has four convolutional layers, aimed at extracting local image features, followed by two fully connected layers and a sigmoid layer for classification. Each layer is followed by a non-linear hyperbolic tangent (tanh) activation function. For classification based on the HOG, LBP, and DEA features, we remove the convolutional layers and feed the features directly to the fully connected layers and then the sigmoid layer for the classification. For the RF, SVM, and KNN classifiers we use the implementation of these classifiers from the Scikit-Learn library (39) in python.

FIGURE 3
www.frontiersin.org

Figure 3. The CNN architecture used in the classification task. Convolutional layers are made up of (5 × 5 × 5) kernels with a stride of (2 × 2 × 2) which reduces the volume by half of the input size on each layer. The first two layers extract 2 feature cubes and the last two layers extract 4 feature cubes each. The fully connected layers have 64 and 32 hidden nodes respectively and the convolutional and fully connected layers are followed by a non-linear hyperbolic tangent (tanh) activation function.

3. Experiments and results

3.1. Patient population and image data

We test our proposed methods on parametric volumes extracted from MR perfusion data from 183 patients with acute ischemic stroke. Details of the image acquisition and preparation are already published by Pinto et al. (40). Our dataset is made up of three parametric information—Tmax volumes which refer to the time taken for the blood flow to reach its peak, relative blood flow (rBF) volumes which refer to the volume of blood passing through a given brain tissue per unit of time, and relative blood volume (rBV) defined as the volume of blood in a given brain tissue relative to an internal control (e.g., normal white matter or an arterial input function). Each volume has a resolution of (0.9, 0.9, and 6.5 mm) and a dimension of (256, 256, and 19) voxels on the sagittal, coronal and axial planes respectively. Ground truth labels are obtained from a trained neuroradiologist, with over ten years of experience, who manually investigates the DSA slides of the associated patient and assigns one of three labels (0-poor, 1-medium, 2-good) to this patient. We use these labels for a 3-class prediction experiment and we also experiment on a risk-stratified nested test where we first predict good - (2) against not good (0, 1) collaterals and then separate the not good class into poor (0) and medium (1) collaterals in a cascaded approach.

3.2. Preprocessing

Our image preprocessing involves two main tasks. First, we make our datasets isotropic by applying a B-spline interpolation to the axial axis since the other two axes have the same spacing leading to volume with a resolution of 0.9 mm on each plane and a new dimension of (256, 256, and 127). This is followed by an extraction of the brain region from the skull using the brain extraction tool (BET) from the ANTS library. The brain extraction is carried out on the Tmax volumes and the resulting mask is then applied to the rBF and rBV volumes.

3.3. Region of interest localization

After the preprocessing step we extract the occluded regions as the region of interest (ROI) using the reinforcement learning architecture described in Section 2.1. We adopt the network architecture from Alansary et al. (33) with modifications proposed in Navarro et al. (35). A stopping criterion of τ = 0.85 is used during training—that is, an intersection over union (IoU) value greater than or equal to 0.85 implies that the region of interest is detected. We perform the ROI detection task on the Tmax volumes since the occluded regions are easier to detect in these volumes. The resulting cube region is then applied on the rBF and rBV volumes to extract the corresponding cubes in these volumes as well. For each volume, we select 20 starting cubes of size (64 × 64 × 64) at random and run the agent till the stopping criterion is reached. We then aggregate the results from the 20 different runs to get the prediction of the final ROI. After getting the region we extract the mirror of the ROI (ROI+M) by reflecting the ROI on the opposite side of the brain and using it as an additional feature. This results in 6 cubes per patient (i.e., two volumes each from Tmax, rBF, and rBV volumes). Qualitative and quantitative results from the region of interest extraction can be found in Table 1 and Figures 4, 5. From the box plots in Figure 5, it is evident that the region of interest detection was more successful in the poor collateral flow classes (class 0 and 1) than in the good collateral flow class. This can be explained by the fact that in cases of good collateral flow, there is a uniform distribution of the Tmax value within the occluded region and its neighborhood making it hard for the RL agent to detect the ROI. From Figure 4 we observe that in most cases the ground truth does not cover the total occluded region [e.g., column (b)] and hence the predicted ROI, though does not completely overlap the ground truth, still contains other parts of the occluded region which is not captured in the ground truth and it is therefore sufficiently accurate for the classification task.

TABLE 1
www.frontiersin.org

Table 1. Quantitative results from the region of interest detection task.

FIGURE 4
www.frontiersin.org

Figure 4. Qualitative results from the ROI detection task. The top row is the axial view of the ground truth (in red) and the prediction (in green). The bottom row is a 3-D visualization of the ground truth cube (in red), the predicted cube (in green), and the intersection between the two (in blue). Column (A) corresponds to the worst prediction in our test set while column (C) refers to the best result in terms of IoU. In column (B), we can observe that though the overlap is not perfect the prediction still contains some part of the occluded region which is not in the ground truth. This implies that though we have poor scores we still have good ROI detection which can be used for the classification task.

FIGURE 5
www.frontiersin.org

Figure 5. Box plots of results from ROI detection task. Left is the intersection over union (IoU) ratio between the prediction and the ground truth over the three classes. Right is the euclidean distance between the predicted center point and the ground truth center point. From the distributions, it is clear that it is easy to detect the ROI in the poor collateral flow class (class 0) compared to the good collateral flow class (class 2). This can be explained by the fact that in good collateral flow cases Tmax shows uniform values in the whole volume.

3.4. Classification

3.4.1. Feature representations

In total three sets of features (DAE, HOG, and LBP) are extracted in addition to the actual extracted cube (ROI) and its mirror cube (ROI+M). We learn features automatically through an unsupervised denoising auto-encoder. The network takes the extracted ROI cubes from the Tmax, rBV, and rBF volumes as three input channels and produces a single channel feature set of size (8 × 8 × 8). We normalize the cubes individually into the range [0, 1] before feeding them to the network.

For HOG features we extract 10 features each for the three parametric volumes and concatenate them into a vector of length 30 for the classification task. Figure 6 shows a sample of the extracted HOG features for a patient for the three input channels. Finally, LBP features are extracted using the method described in Section 2.2. Here we combine all the three channels and run the histogram over the three channels which results in a 927 feature vector as explained in Section 2.2. Figure 6 shows a sample of the extracted LBP features from our dataset.

FIGURE 6
www.frontiersin.org

Figure 6. Sample of extracted features using Local Binary Pattern (LBP) on the left and Histogram of Oriented Gradients (HOG) on the right. Bar heights from LBP features represent the frequency of a given pattern and the position on the x-axis is the decimal representation of the binary pattern. Arrow directions in HOG features are the gradient vectors and the length of the arrows represents the frequency of the given gradient. The LBP features show a uniform distribution of the extracted patterns with no dominant pattern. The HOG features on the other hand show evidence of high gradients in the extracted region of interest.

3.4.2. Classifier training

We handle the collateral flow classification through two main approaches—a three-class multi-label classification task where we predict three labels in one step, and a two-step cascaded approach where we predict a binary label of classes (0, 1) against 2 in the first step and separate the class 0 from 1 in the second step. We implement our CNN architecture using the Keras library (41) in python with TensorFlow as the backend. Random forest, support vector machine, and K-nearest neighbor classifiers were implemented using the scikit-learn library (39) in python. We set up our experiments as follows:

CNN classifier: For the CNN classifier, we use a weighted categorical cross-entropy with a weight of 1|k| for each class k in the training set. A stochastic gradient descent optimizer with a learning rate of 0.001, decay of 1e−6, and momentum of 0.9 is used to fine-tune the network parameters at 20 epochs.

K-Nearest neighbor classifier: We conduct preliminary a experiment with a grid search to know which parameters will work best. For our final experiment, we use k = 3 neighbors with uniform weights, a leaf size of 30, and the Minkowski metric.

Random forest classifier: After the initial grid search experiment, we implement the classifier with 200 estimators, and the Gini impurity function is used to measure the quality of a split.

Support vector machine classifier: We use a regularization parameter C = 10, a third-degree polynomial kernel, a balanced class weight, and a tolerance of 1e−3 for the stopping criterion.

3.4.3. Classification results

We test different combinations of the feature sets extracted in the previous experiments and classifiers discussed in a preliminary experiment and present the results in Table 2. Due to limitations in the size of the dataset, we adopt a 5-fold cross-validation approach in a preliminary experiment instead of a training-validation-test splitting approach and report the average scores over the accuracy in the individual validations. In the preliminary experiments (results in Table 2), we use the manually annotated ROI and not the ROI predicted from the proposed reinforcement learning. We later, in a follow-up experiment, compare the performance of the proposed CNN on manually annotated ROI and the predicted ROI (results in Table 3).

TABLE 2
www.frontiersin.org

Table 2. Results from preliminary experiments on collateral flow grading.

TABLE 3
www.frontiersin.org

Table 3. Results from the experiment on collateral flow grading using only ROI data on our proposed cascaded CNN.

The results in Table 2 show that the region of interest extraction step helps improve the results in all classification methods. This can be verified by comparing the performance from the full image (RAW column) with the performance of the region of interest (ROI column) in Table 2. Also by adding the mirror of the occluded region to the extracted ROI (ROI+M) we achieve improved results in most of the classifiers with performance falling in classifiers like KNN and SVM due to the increase in the dimension of data introduced by the mirror of the ROI. The cascaded method shows higher accuracy in almost all the classifier-feature combinations when compared to the direct three-class prediction. This can be explained by the distribution of classes in the dataset. That is, for the cascaded approach we have fairly balanced data when we combine poor and moderate flow against good collateral flow which is not the case with the direct three-class multi-label prediction approach. It, therefore, suggests that in cases where we have highly imbalanced class distributions a multi-label classification might perform poorly. The overall performance of CNN is better than the other machine learning classifiers and can be explained by the fact that the convolutional layers of the CNN architecture extract features while paying attention to the class of the input data. This makes the feature extraction process more efficient than the other feature extraction schemes which have no knowledge of the underlying label of the input data at the time of extracting the features. Again CNN with only ROI data performs slightly better than with the mirror of the ROI (72 vs. 70% in Table 2) and this can also be explained by the fact that the CNN used in our experiments is fairly shallow and hence could not handle the additional feature dimensions introduced by the mirrored images.

Based on the results of the preliminary experiment, we further probe into the training of the proposed CNN classifier with the ROI data. In this experiment, we split the data into training and testing sets. The test set is made up of 50 volumes randomly selected with reference to the ratio of class count in the entire dataset. We make use of both the manually annotated ROI and the automated ROI from our proposed Reinforcement Learning approach during training. We finally evaluate the trained models on the automated ROI and compare it with the same network trained and evaluated solely on the manually annotated ROI data. Table 3 shows the result of this experiment.

The results in Table 3 from our follow-up experiment show that the automated ROI from the proposed Reinforcement Learning approach is comparable to the manually detected ROI in terms of predicting collateral flow (2% drop in accuracy which represents one out of the 50 patients in the test set). This is crucial in automating the whole collateral flow prediction workflow in a clinical setting.

4. Summary and conclusion

In this work, we present a deep learning approach toward grading collateral flow in ischemic stroke patients based on parametric information extracted from MR perfusion data. We start by extracting regions of interest using deep reinforcement learning. We then learn denoising auto-encoder features and modern implementation of 3-D HOG and LBP features. We proceed to the actual classification task using a combination of the extracted features and CNN, random forest, K-nearest neighbor, and support vector machine classifiers.

Our experiments show that the rich information on blood flow visible from MRI perfusion can be used to predict collateral flow in a similar manner to DSA images which are invasive in nature. Region of interest detection with reinforcement learning is successful to an acceptable level and can be used as a guide to estimate the region in the brain which requires more attention. It is evident that high class imbalance can be a major challenge in the collateral flow grading task and many similar works. We however show that for datasets with high class imbalance, a two-step cascaded classification approach performs better than a one-time multi-label classification method. It is also evident from our results that a direct CNN classifier is able to extract relevant features from the region of interest and has an advantage over classical machine learning classifiers like RF, KNN, and SVM that depend on handcrafted features like HOG and LBP.

Collateral flow grading is an essential clinical procedure in the treatment of ischemic stroke patients. We have presented a framework for automating the process in clinical setup and have achieved promising results given our limited dataset. For the proposed framework to be clinically useful there is the need for further tests with possibly more data from multiple stroke centers. The grading can also be customized for specific patient groups for example providing information about age group, gender, and other biographical and historical information of patients as an additional feature can help improve the result of the framework.

Data availability statement

Data is currently private, pending approval from the source of the data. Requests to access these datasets should be directed to giles.tetteh@tum.de.

Author contributions

GT and BM contributed to the conception and design of the study. GT, FN, and JP contributed to the experiments and computer codes. RM, JK, and RW organized the database. GT wrote the first draft of the manuscript. BM, JP, and JK contributed to the final manuscript. BM, CZ, JK, and RW handled the administrative tasks. All authors contributed to manuscript revision, read, and approved the submitted version.

Funding

GT and BM acknowledge support by the Deutsche Forschungsgemeinschatf (DFG) (ME 3511/5-1).

Acknowledgments

BM acknowledges support from the Helmut-Horten-Foundation.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Faber J, Chilian W, Deindl E, van Royen N, Simons M. A brief etymology of the collateral circulation. Arterioscler Thromb Vasc Biol. (2014) 9:1854–9. doi: 10.1161/ATVBAHA.114.303929

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Liu L, Go T, Leng X, Pu Y, Huang LA, Xu A, et al. Guidelines for evaluation and management of cerebral collateral circulation in ischaemic stroke 2017. Stroke Vasc Neurol. (2018) 3:svn-2017. doi: 10.1136/svn-2017-000135

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Ginsberg M. The cerebral collateral circulation: Relevance to pathophysiology and treatment of stroke. Neuropharmacology. (2017) 134:280–92. doi: 10.1016/j.neuropharm.2017.08.003

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Ben Hassen W, Malley C, Boulouis G, Claren?on Claren?on F, Bartolini B, Bourcier R, et al. Inter- and intraobserver reliability for angiographic leptomeningeal collateral flow assessment by the American Society of Interventional and Therapeutic Neuroradiology/Society of Interventional Radiology (ASITN/SIR) scale. J Neurointerv Surg. (2018) 11:338–41. doi: 10.1136/neurintsurg-2018-014185

PubMed Abstract | CrossRef Full Text | Google Scholar

5. McVerry F, Liebeskind D, Muir K. Systematic review of methods for assessing leptomeningeal collateral flow. AJNR Am J Neuroradiol. (2011) 33:576–82. doi: 10.3174/ajnr.A2794

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Martinon E, Lefevre P, Thouant P, Osseby G, Ricolfi F, Chavent A. Collateral circulation in acute stroke: Assessing methods and impact: A literature review. J Neuroradiol. (2014) 41:97–107. doi: 10.1016/j.neurad.2014.02.001

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Sheth SA, Liebeskind DS. Imaging evaluation of collaterals in the brain: physiology and clinical translation. Curr Radiol Rep. (2014) 29:1–9. doi: 10.1007/s40134-013-0029-5

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Jansen IGH, Berkhemer OA, Yoo AJ, Vos JA, Lycklama A, Nijeholt GJ, et al. Comparison of CTA- and DSA-based collateral flow assessment in patients with anterior circulation stroke. AJNR Am J Neuroradiol. (2016) 37:2037–42. doi: 10.3174/ajnr.A4878

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Muehlen I, Kloska SP, Gölitz P, Hölter P, Breuer L, Ditt H, et al. Noninvasive collateral flow velocity imaging in acute ischemic stroke: intraindividual comparison of 4D-CT angiography with digital subtraction angiography. In: RöFo-Fortschritte auf dem Gebiet der Röntgenstrahlen und der bildgebenden Verfahren. Georg Thieme Verlag KG. (2019). Vol. 191, p. 827-835. doi: 10.1055/a-0825-6660

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Kauw F, Dankbaar JW, Martin BW, Ding VY, Boothroyd DB, van Ommen F, et al. Collateral status in ischemic stroke: a comparison of computed tomography angiography, computed tomography perfusion, and digital subtraction angiography. J Comput Assist Tomogr. (2020) 44:984–92. doi: 10.1097/RCT.0000000000001090

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Jung S, Wiest R, Gralla J, McKinley R, Mattle H, Liebeskind D. Relevance of the cerebral collateral circulation in ischaemic stroke: Time is brain, but collaterals set the pace. Swiss Med Wkly. (2017) 147:w14538. doi: 10.4414/smw.2017.14538

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Maija R. Collateral blood supply in acute stroke. (2015). doi: 10.1594/ecr2015/C-1460

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Bang OY, Saver J, Kim S, Kim GM, Chung CS, Ovbiagele B, et al. Collateral flow predicts response to endovascular therapy for acute ischemic stroke. Stroke. (2011) 42:693–9. doi: 10.1161/STROKEAHA.110.595256

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Bang OY, Saver J, Buck B, Alger JR, Starkman S, Ovbiagele B, et al. Impact of collateral flow on tissue fate in acute ischaemic stroke. J Neurol Neurosurg Psychiatry. (2008) 79:625–9. doi: 10.1136/jnnp.2007.132100

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Mansour O. Assessment of collateral pathways in acute ischemic cerebrovascular stroke using a mansour grading scale; a new scale, a pilot study. Internet J Internal Med. (2013) 3:1–10.

Google Scholar

16. Maas M, Lev M, Ay H, Singhal A, Greer D, Smith W, et al. Collateral vessels on CT angiography predict outcome in acute ischemic stroke. Stroke. (2009) 40:3001–5. doi: 10.1161/STROKEAHA.109.552513

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Tan I, Demchuk A, Hopyan J, Zhang L, Gladstone D, Wong K, et al. CT angiography clot burden score and collateral score: correlation with clinical and radiologic outcomes in acute middle cerebral artery infarct. AJNR Am J Neuroradiol. (2009) 30:525–31. doi: 10.3174/ajnr.A1408

PubMed Abstract | CrossRef Full Text | Google Scholar

18. McHugh ML. Interrater reliability: the kappa statistic. Croatian Soc Med Biochem Lab Med. (2012) 22:276–82. doi: 10.11613/BM.2012.031

CrossRef Full Text | Google Scholar

19. Kersten-Oertal M, Alamer A, Fonov V, Lo BWY, Tampieri D, Collin DL. Towards a computed collateral circulation score in ischemic stroke. arXiv [Preprint]. (2020). arXiv: 2001.07169. doi: 10.48550/ARXIV.2001.07169

CrossRef Full Text | Google Scholar

20. Grunwald IQ, Kulikovski J, Reith W, Gerry S, Namias R, Politi M, et al. Collateral automation for triage in stroke: evaluating automated scoring of collaterals in acute stroke on computed tomography scans. Cerebrov Dis. (2019) 47:1–6. doi: 10.1159/000500076

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Pauli J. Learning-Based Robot Vision, Principles and Applications. Berlin: Springer Science & Business Media (2001). doi: 10.1007/3-540-45124-2

CrossRef Full Text | Google Scholar

22. DeSouza G, Kak A. Vision for mobile robot navigation: a survey. IEEE Trans Pattern Anal Mach Intell. (2002) 24:237–67. doi: 10.1109/34.982903

CrossRef Full Text | Google Scholar

23. Bonin-Font F, Ortiz A, Oliver G. Visual navigation for mobile robots: a survey. J Intell Robot Syst. (2008) 53:263–96. doi: 10.1007/s10846-008-9235-4

CrossRef Full Text | Google Scholar

24. Peng J, Bhanu B. Delayed reinforcement learning for closed-loop object recognition. In: Proceedings of 13th International Conference on Pattern Recognition. IEEE. (1996). p. 310–314.

Google Scholar

25. Peng J, Bhanu B. Closed-loop object recognition using reinforcement learning. IEEE Trans Pattern Anal Mach Intell. (1998) 20:139–54. doi: 10.1109/34.659932

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Taylor G, Wolf C. Reinforcement learning for parameter control of text detection in images from video sequences. In: Proceedings 2004 International Conference on Information and Communication Technologies: From Theory to Applications. (2004). p. 517–518.

Google Scholar

27. Sahba F, Tizhoosh HR, Salama MM. A reinforcement learning framework for medical image segmentation. In: The 2006 IEEE international joint conference on neural network proceedings. IEEE (2006). p. 511–517. doi: 10.1109/IJCNN.2006.246725

CrossRef Full Text | Google Scholar

28. Sahba F, Tizhoosh HR, Salama MM. Application of opposition-based reinforcement learning in image segmentation. In: 2007 IEEE Symposium on Computational Intelligence in Image and Signal Processing. IEEE (2007). p. 246–251. doi: 10.1109/CIISP.2007.369176

CrossRef Full Text | Google Scholar

29. Shokri M, Tizhoosh HR. Using reinforcement learning for image thresholding. In: CCECE 2003-Canadian Conference on Electrical and Computer Engineering. IEEE (2003). p. 1231–1234.

PubMed Abstract | Google Scholar

30. Sahba F, Tizhoosh HR, Salama MM. Using reinforcement learning for filter fusion in image enhancement. In: IASTED International Conference on Computational Intelligence. (2005). p. 262–266.

Google Scholar

31. Tizhoosh H, Taylor G. Reinforced contrast adaptation. Int J Image Graphics. (2006) 6:377–92. doi: 10.1142/S0219467806002379

CrossRef Full Text | Google Scholar

32. Netto SM, Leite VR, Silva AC, de Paiva AC, de Almeida Neto A. Application on reinforcement learning for diagnosis based on medical image. In: Reinforcement Learning. IntechOpen (2008) 379.

Google Scholar

33. Alansary A, Oktay O, Li Y, Folgoc L, Hou B, Vaillant G, et al. Evaluating reinforcement learning agents for anatomical landmark detection. Med Image Anal. (2019) 53:156–64. doi: 10.1016/j.media.2019.02.007

PubMed Abstract | CrossRef Full Text | Google Scholar

34. Mnih V, Kavukcuoglu K, Silver D, Rusu A, Veness J, Bellemare M, et al. Human-level control through deep reinforcement learning. Nature. (2015) 518:529–33. doi: 10.1038/nature14236

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Navarro F, Sekuboyina A, Waldmannstetter D, Peekan JC, Combs SE, Menze BH. Deep Reinforcement Learning for Organ Localization in CT. Proc Mach Learn Res. (2020) 121:544–54.

Google Scholar

36. Milletari F, Navab N, Ahmadi SA. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In: 2016 Fourth International Conference on 3D vision (3DV). IEEE (2016). p. 565–571. doi: 10.1109/3DV.2016.79

CrossRef Full Text | Google Scholar

37. Heikklä M, Pietikäinen M. A texture-based method for modeling the background and detection moving objects. IEEE Trans Pattern Anal Mach Intell. (2006) 28:657–62. doi: 10.1109/TPAMI.2006.68

PubMed Abstract | CrossRef Full Text | Google Scholar

38. Klaser A, Marszaäek M, Schmid C. A spatio-temporal descriptor based on 3d-gradients. In: BMVC 2008-19th British Machine Vision Conference. British Machine Vision Association (2008). p. 275–1. doi: 10.5244/C.22.99

CrossRef Full Text | Google Scholar

39. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-learn: Machine learning in python. J Mach Learn Res. (2011) 12:2825–30.

Google Scholar

40. Pinto A, Amorim J, Hakim A, Alves V, Reyes M, Silva CA. Prediction of stroke lesion at 90-day follow-up by fusing raw DSC-MRI with parametric maps using deep learning. IEEE Access. (2021) 9:26260–70. doi: 10.1109/ACCESS.2021.3058297

CrossRef Full Text | Google Scholar

41. Chollet F. Keras: GitHub (2015). Available online at: https://github.com/fchollet/keras (accessed August 31, 2022).

Google Scholar

Keywords: collateral flow, radiomics, perfusion, reinforcement learning, image descriptors, angiography, auto-encoder, deep learning

Citation: Tetteh G, Navarro F, Meier R, Kaesmacher J, Paetzold JC, Kirschke JS, Zimmer C, Wiest R and Menze BH (2023) A deep learning approach to predict collateral flow in stroke patients using radiomic features from perfusion images. Front. Neurol. 14:1039693. doi: 10.3389/fneur.2023.1039693

Received: 08 September 2022; Accepted: 02 February 2023;
Published: 21 February 2023.

Edited by:

Jean-Claude Baron, University of Cambridge, United Kingdom

Reviewed by:

Anna Podlasek, University of Dundee, United Kingdom
Arvind Bambhroliya, University of Texas Health Science Center at Houston, United States
Jennifer Soun, University of California, Irvine, United States

Copyright © 2023 Tetteh, Navarro, Meier, Kaesmacher, Paetzold, Kirschke, Zimmer, Wiest and Menze. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Giles Tetteh, yes giles.tetteh@tum.de

Present address: Johannes C. Paetzold, Department of Computing, Imperial College London, London, United Kingdom

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.