Abstract
The mechanical contraction of the pumping heart is driven by electrical excitation waves running across the heart muscle due to the excitable electrophysiology of heart cells. With cardiac arrhythmias these waves turn into stable or chaotic spiral waves (also called rotors) whose observation in the heart is very challenging. While mechanical motion can be measured in 3D using ultrasound, electrical activity can (so far) not be measured directly within the muscle and with limited resolution on the heart surface, only. To bridge the gap between measurable and not measurable quantities we use two approaches from machine learning, echo state networks and convolutional autoencoders, to solve two relevant data modelling tasks in cardiac dynamics: Recovering excitation patterns from noisy, blurred or undersampled observations and reconstructing complex electrical excitation waves from mechanical deformation. For the synthetic data sets used to evaluate both methods we obtained satisfying solutions with echo state networks and good results with convolutional autoencoders, both clearly indicating that the data reconstruction tasks can in principle be solved by means of machine learning.
1 Introduction
Cardiac arrhythmias, such as ventricular or atrial fibrillation, are electro-mechanical dysfunctions of the heart that are associated with complex, chaotic spatio-temporal excitation waves within the heart muscle resulting in incoherent mechanical contraction and a significant loss of pump function [1–3]. Ventricular fibrillation (VF) is the most common deadly manifestation of a cardiac arrhythmia and requires immediate defibrillation using high-energy electric shocks. Atrial fibrillation (AF) is the most common form of a cardiac arrhythmia, affecting 33 million patients worldwide [62]. While not immediately life-threatening, AF is considered to be responsible for 15% of strokes if left untreated [63, 64]. The structural substrate and functional mechanisms that underlie the onset and perpetuation of VF and AF are not fully understood. It is generally agreed that imaging of the cardiac electrical and mechanical function is key to an improved mechanistic understanding of cardiac disease and the development of novel diagnosis and therapy. This has motivated the development of non-invasive and invasive electrophysiological measurement and imaging modalities. Electrical activity of the heart can (so far) non-invasively be measured on its surface, only. Direct measurements can be made in vivo inside the heart using so-called basket catheters with typically 64 electrodes or in ex-vivo experiments, where an extracted heart in a Langendorff perfusion set-up is kept beating and the cell membrane voltage on the epicardial surface is made visible using fluorescent dyes (a method also known as optical mapping) [4]. A method for indirect observation of electrical excitation waves is ECG imaging where an array of EEG-electrodes is placed on the body surface and an (ill-posed) inverse problem is solved to estimate the potential on the surface of the heart. Mechanical contraction and deformation of the heart tissue can be studied in full 3D using ultrasound, in 2D using real-time MRT [5] or (using optical mapping) by motion tracking in Langendorff experiments.
The reconstruction of patterns of action potential wave propagation in cardiac tissue from ultrasound has been introduced by Otani et al. [6, 7]. They proposed to use ultrasound to visualize the propagation of these waves through the mechanical deformations they induce and to reconstruct action potential-induced active stress from the deformation. Provost et al. [8] introduced electromechanical wave imaging to map the mechanical deformation of cardiac tissue at high temporal and spatial resolutions. The observed deformations resulting from the electrical activation were found to be closely correlated with electrical activation sequences. The cardiac excitation-contraction-coupling (ECC) [9] has also been studied in optical mapping experiments in Langendorff-perfused isolated hearts [10–12]. Using electromechanical optical mapping [12], it was shown that during ventricular tachyarrhythmias electrical rotors introduce corresponding rotating mechanical waves. These co-existing electro-mechanical rotors were observed on the epicardial surface of isolated Langendorff-perfused intact pig and rabbit hearts using optical mapping [13]. Using high-resolution ultrasound, these mechanical rotors were also observed inside the ventricular wall during ventricular tachycardia and fibrillation [13].
All these measurement modalities are limited, in particular those suitable for in vivo applications. Measurements with basket catheters are effectively undersampling the spatio-temporal wave pattern. Inverse ECGs suffer from ill-posedness and require regularization that may lead to loss of spatial resolution and blurring. Limited spatial resolution is also an issue with ultrasound measurements, but they are currently the only way to “look inside” the heart, albeit measuring only mechanical motion. Electrical excitation waves inside the heart muscle are so far not accessible by any measurement modality available.
These limitations motivated the search for algorithms to reconstruct electro-mechanical wave dynamics in cardiac tissue from measurable quantities. Berg et al. [14] devised synchronization-based system identification of extended excitable media, in which model parameters are estimated by minimizing the synchronization error. Using this approach, Lebert and Christoph [15] demonstrated that electro-mechanic wave dynamics of excitable-deformable media can be recovered from a limited set of observables using a synchronization-based data assimilation approach. Hoffman et al. reconstructed electrical wave dynamics using ensemble Kalman filters [16, 17]. In another approach, it was shown that echo state networks [18] and deep convolutional neural networks [19, 20] provide excellent cross estimation results for different variables of a mathematical model describing complex electrical excitation waves during cardiac arrhythmias. Following this approach, Christoph and Lebert [21] demonstrated the reconstruction of electrical excitation and active stress from deformation using a simulated deformable excitable medium. To continue this research and to address the general challenge of missing or impaired observations we consider in this article two tasks: (i) recovering electrical excitation patterns from noisy, blurred or undersampled observations and (ii) reconstructing electrical excitation waves from mechanical deformation. To solve the corresponding data processing and cross-prediction tasks two machine learning methods are employed and evaluated: echo state networks and convolutional autoencoders. Both algorithms are applied to synthetical data generated by prototypical models for electrophysiology and electromechanical coupling.
2 Methods
In this section we will first introduce in Sections 2.1 and 2.2 the mathematical models describing cardiac dynamics which were used to generate the example data for the two tasks to be solved: (i) recovering electrical wave pattern from impaired observations and (ii) cross-predicting electrical excitation from mechanical deformation. Then in Section 2.3 both machine learning methods used for solving these tasks, echo state networks (Section 2.3.1) and convolutional autoencoders (Section 2.3.2), will be briefly introduced.
2.1 Recovering Complex Spatio-Temporal Wave Patterns From Impaired Observations
For motivating, illustrating, and evaluating the employed methods for dealing with incomplete or distorted observations we shall use spatio-temporal time series generated with the Bueno-Orovio-Cherry-Fenton (BOCF) model [22] describing complex electrical excitation patterns in the heart during cardiac arrhythmias. The BOCF model is a set of partial differential equations (PDEs) with four variables and will be introduced in Section 2.1.1. In Section 2.1.2 a formal description of the data recovery tasks will be given.
2.1.1 Bueno–Orovio–Cherry–Fenton Model
Cardiac dynamics is controlled by electrical excitation waves triggering mechanical contractions of the heart. In the case of cardiac arrhythmias like lethal ventricular fibrillation, wave break-up and complex chaotic wave patterns occur resulting in significantly reduced pump performance of the heart. From the broad range of mathematical models describing this spatio-temporal dynamics [23] we chose the Bueno–Orovio–Cherry–Fenton (BOCF) model [22] to generate spatio-temporal time series that are used as a benchmark to validate our approaches for reconstructing complex wave patterns in excitable media from incomplete data. The BOCF model consists of four system variables whose evolution is given by four (partial) differential equations
The variable u represents the continuum limit representation of the membrane voltage of cardiac cells and the variables v, w, and s are gating variables controlling ionic transmembrane currents , and given by the equations
Here denotes the Heaviside function and the currents depend on the following seven voltage controlled variables
For simulating the dynamics we used the set of parameters given in Table 1 for which the BOCF model was found [22] to exhibit excitation wave dynamics similar to the Ten Tusscher–Noble–Noble–Panfilov (TNNP) model [24] describing human heart tissue.
TABLE 1
| 0 | 1150 | 0.11 | 2.7342 | 3 | 6 | 6 | |||||||
| 1.58 | 1.4506 | 70 | 20 | 43 | 0.2 | 2.8723 | |||||||
| 0.3 | 0.07 | 60 | 280 | 2.0994 | 0.94 | 0.015 | |||||||
| 0.9087 | 0.015 | 65 | 0.006 | 0.03 | 2 | 0.65 |
TNNP model parameter values for the BOCF model [22].
Typical snapshots of the four variables during a chaotic evolution are shown in Figure 1. The spatio-temporal chaotic dynamics of this system is actually transient chaos whose lifetime grows exponentially with system size [25, 26]. To obtain chaotic dynamics with a sufficiently long lifetime the system has been simulated on a domain of grid points with a grid constant of space units and a diffusion constant . Furthermore, an explicit Euler stepping in time with , a 5 point approximation of the Laplace operator, and no-flux boundary conditions were used for solving the PDEs.
FIGURE 1
2.1.2 Reconstruction Tasks
Experimental measurements of the dynamics of a system of interest often allow only the observation of some state variables (e.g., the membrane voltage) and may provide only incomplete or distorted information about the measured observable. Typical limitations are (additive) measurement noise and low-spatial resolution (due to the experimental conditions and/or the available hardware). Formally, measurements impaired due to noise, blurring or undersampling can be described as follows: Let
be the measured data (here: snapshots of the field
u) where
rand
cspecify the two spatial dimensions. Each sample
with
corresponds to a true system output
that is assumed to be known only during the training phase in terms of a training set
. Note that with coarse graining
and
. The task is to predict the true system output
from impaired observations
which belong to one of the following three cases:
- 1.
Noisy data: To add noise each element of is replaced with probability p by 0 or 1 drawn from a Bernoulli distribution (note that in our case is given by the variable u of the BOCF model which has a range of ). To simulate different levels of noise different probabilities are used to generate noisy data sets . In the following p is called the noise level.
- 2.
Blurred data: Date with reduced spatial resolution are obtained as Fourier low-pass filtered data where and denote the Fourier transform and its inverse, respectively, and is a projection where frequencies outside a radius (Manhattan distance) centered at frequency zero are set to zero.
- 3.
Undersampled data: To generate undersampled date is down-sampled with and by accessing every 2i-th value of , where .
Figure 2 shows examples of the three types of impaired observations.
FIGURE 2
2.2 Predicting Electrical Excitation From Mechanical Contraction
To learn the relation between mechanical deformation and electrical excitation inverse modelling data were generated by a conceptual electro-mechanical model consisting of an Aliev–Panfilov model describing the electrical activity and a driven mass-spring-system [15].
2.2.1 Aliev–Panfilov Model
Specifically developed to mimic cardiac action potentials in the myocardium, the Aliev–Panfilov model is a modification of the FitzHugh–Nagumo model, which reproduces the characteristic shape of electric pulses occurring in the heart [27]. It is given by a set of two differential equations,in which u and v are the normalized membrane voltage and the recovery variable, respectively, and a, b and k are model parameters. The term accounts for the diffusion, in which the tensor can be used to model anisotropies in the myocardial tissue. In addition, the term is introduced to adjust the shape of the restitution curve by modulating the parameters and . The computational advantage of the Aliev–Panfilov model lies in its simplicity over other ion-flow-based models which allows shorter runtimes and combined with the elastomechanical model, keeps computational costs fairly reasonable. For this reason, the Aliev–Panfilov model was chosen for generating synthetic data from complex chaotic electromechnical wave dynamics.
Within the heart muscle, the myocardium, cells contract upon electrical excitation through a passing action potential. At this point it is important to note that muscle fibre contracts along its principal orientation which has to be considered during the implementation of the mechanical part of the simulation. To couple the mechanical contraction of the muscle fibre to electrical excitation of a cell, as an extension to the Aliev–Panfilov model the active stress was introduced by Nash and Panfilov [28] which leads to contraction in the principal orientation of the muscle fibre. The change of the active stress is described bywhere controls the strength of the build-up of active stress. The term regulates the influence of u on for large u. In our simulations we use a smooth function introduced by Göktepe and Kuhl [29] given by
Here, controls the steepness of the transition between and and denotes the potential threshold for the activation of the active stress, with to achieve a physiological time course [30].
2.2.2 Mass-Spring Damper System
The elasto-mechanical properties of the cardiac muscle fibre were implemented using a modified two-dimensional mass-spring damper system [31]. For the current study the mass-spring system was implemented in two dimensions because this allows shorter runtimes and primarily serves as a proof-of-principle for the evaluated reconstruction approach. In its two-dimensional form this mechanical model might correspond best to a cut-out of the atrium’s wall, since there the muscle tissue is less than 4 mm thick. However, this mass-spring system can easily be expanded to three dimensions (see [15]).
Placed on a regular lattice, one mechanical cell is made up of four particles at the corners connected by structural springs and two sets of orthogonal springs connecting the centre of mass to each side of the cell (see Figure 3). The springs in the middle of the cell are called axial springs, of which one is made to be active (red). Here it is important to point out that one cell in the electrical model corresponds to one cell in the mechanical mass-spring system. For setting the fibre orientation through the active axial spring, the orientation parameter has been introduced, with which the four points of attachment can be computed easily. This parameter can be set individually for each cell, so that various fibre orientations can be modelled.
FIGURE 3
Using the forces from the passive spring and the active spring are obtained as.
Here , and , denote the resting lengths and spring constants of the passive and active spring, respectively. From Eq. (9) it can be seen that, upon a rise in active stress from Eq. (6), the active spring contracts and an inward force is generated. The parameter represents a scaling factor to modulate the influence of the active stress. Through the orientation parameter the forces from the active and passive spring can be redistributed to the corresponding particles at the corners. For example for , the force on would be and on it would amount to .
In addition, the mechanical grid is held together by structural forces between the corner particles, which can be computed usingwith being the resting length between particle and .
Finally, with all the above forces acting on particle with mass , its motion is determined according towith the sum of all relevant springs pulling or pushing the particle. The damping constant ν sets the strength of the damping to increase the stability of the mechanical system as a whole.
The area of each cell was calculated with a simple formula for a general quadrilateral using the positions of its four corners. As a measure of contraction, the relative change of areahas been used. The numerical algorithm for solving the full set of electro-mechanical ODEs is summarized in the Appendix.
2.2.3 Reconstruction Task
The inverse modelling data are generated by forward modelling using the output of Equations (4) and (13). The task is to train an ESN or CAE to approximate . To fulfill this task we use the membrane voltages and the local deformations at all grid points sampled at times . The training data set thus consists of snapshots and of the relative mechanical deformation and the membrane voltage , respectively, and we aim at approximating with .
2.3 Machine Learning Methods
In this section we will introduce the two machine learning approaches, echo state networks (ESN) [32] and convolutional autoencoders (CAE) [33], that will be applied to solve the reconstruction tasks defined in Section 2.1.2.
2.3.1 Echo State Network
Echo state networks have been introduced in 2001 by Jaeger [32] as a simplified type of recurrent neural network, in which the weights describing the strength of the connections within the network are fixed. In its general composition an ESN subdivides into three sections [32], as illustrated in Figure 4. First of all, there is the input layer into which the input signal and a constant bias are fed. Secondly, the intermediate reservoir consists of N nonlinear units and its state is given by . And lastly, the output layer provides the output signal . Here, n denotes the discrete time steps .
FIGURE 4
The concatenated bias-input vector is fed into the reservoir through the input matrix . Inside the reservoir connections are given by the weight matrix , where N is the reservoir size. Together with the input matrix it is possible to determine the state of the reservoir at time n through the update rulein which denotes a concatenated vector. The input bias , as well as the later introduced output bias were both set to 1 in the following. The parameter in Eq. (14) represents the leaking rate which controls how much of a neuron’s activation is carried over to the next time step and can be used as a parameter to enhance predictions. As for the transfer function we use and the network dynamics used has no feedback loop. Only the weights providing the output signalare adapted during the training process by minimizing the cost function [34]where Tr denotes the trace of a matrix and λ controls the impact of the regularization term that prevents overfitting [35]. The final output matrix is given by the minimum of the cost function at where and are matrices whose columns are given by the vectors and , respectively.
Both matrices and , are initialized with random values from the interval . Since in experiments it turned out that more diverse dynamics could be modelled using networks in which only a small percentage of weights inside the reservoir remained non-zero [32], the weight matrix is made sparse with only a portion of its values remaining non-zero. Furthermore, it is scaled by a factor where denotes here the largest eigenvalue of and ρ is a hyperparameter for optimizing the performance (by ensuring the so-called echo state property [36]). To reduce the probability of drawing an dysfunctional set of matrix entries the randomly generated matrices and were selected from four different realisations. To optimize the performance of the ESN five hyperparameters ( and λ) are tuned.
Reservoir computing using ESNs for predicting chaotic dynamics has already been demonstrated in 2004 by Jaeger and Haas [37]. Since then many studies appeared analyzing and optimizing this approach (see, for example [38–44], and references cited therein). In particular, it has been pointed out how reservoir computing exploits generalized synchronization of uni-directionally coupled systems [45, 46].
Recently, applications of ESNs to spatio-temporal time series have been presented [18, 47] employing many networks operating in parallel at different spatial locations based on the concept of (reconstructed) local states [48]. In particular, using this mode of reservoir computing it was possible to perform a cross-prediction between the four different variables of the BOCF model [18]. Therefore, for the current task of reconstructing data from impaired observations we build on the previous ESN design and modelling procedure. For each pixel an ESN is trained receiving input from neighboring pixels, only, representing the local state at the location of the reference pixel as illustrated in Figure 5. This design introduces two new hyperparameters σ and to the default ESN, where σ is the size of the stencil to define the local state and specifies the spatial distance of adjacent pixels included in the local state. Optimal values for all hyperparameters are determined by a grid search.
FIGURE 5
2.3.2 Convolutional Autoencoder
A convolutional autoencoder [
33] is a special architecture of a feed forward network (FFN) with convolutional layers similar to convolutional neural networks (CNNs) [
49]. Generally a CAE learns a representation of the training set
with the purpose of dimensionality reduction. For each pair
the CAE is trained to perform a nonlinear transformation from the input representation of
to the output representation of
. Like CNNs a CAE is a partially locally connected feed forward network, which is typically composed of the following layers:
Convolutional layers: Convolution of the input by a kernel sliding over the input. The number of rows and columns of the kernel are hyper-parameters, in this work they are set to be .
Batch normalization layer: Normalization of the activations of the previous layer during training and for each batch. Batch normalization allows the use of higher learning rates, being computationally more efficient, and also acts as a regularizer [50].
Leaky ReLU [51] layer: Leaky version of a rectified linear unit (ReLU) [52], such that:
Max pooling layer: Sample-based operation for discretization based on a kernel that slides over the input like the convolutional operator but only the maximum value of the kernel is passed to the next layer. Width and height of the kernel are hyper-parameters (in this work . In contrast to the convolutional layer a pooling layer is not trainable.
Dropout layer: Regularization method to prevent overfitting where during training some weights are set randomly to zero [53]. In this work the probability of setting the weights to zero is 0.05.
The eponymous part of the CAEs are the convolutional layers, a convolution of with a kernel , where , is given by:with . If or exceeds the range of A zero-padding is applied [54].
In this work two architectures are used. The first one employed to reconstruct the data from noisy, blurred and inverse modelling data is illustrated in Figure 6. The architecture is the same for the tasks in Section 2.1.2 and Section 2.2.3 but the sizes of and are different, with and , respectively. Due to the smaller input size, the data for the inverse modelling reconstruction is transformed into a latent space with the size of . The second architecture is sketched in Figure 7 and deals with the undersampled data reconstruction.
FIGURE 6
FIGURE 7
3 Results
In the following both machine learning methods will be applied to two tasks: (i) Reconstructing electrical excitation waves from noisy blurred and under sampled data (Section 3.1) and (ii) Predicting electrical excitation from mechanical contraction (Section 3.2).
3.1 Recovering Complex Spatio-Temporal Wave Patterns From Impaired Observations
To benchmark both reconstruction methods, using ESNs and CAEs, we use time series generated by the BOCF model introduced in Section 2.1.1. The same data were used for both methods, consisting of 5,002 samples in the training data set, 2,501 samples in the validation data set, and 2,497 samples in the test data set. The sampling time of all time series equalled . We considered nine cases of noisy data (with different noise levels), ten cases of (differently) blurred data and seven examples of (spatially) undersampled time series.
For the implemention of the ESN we used the software package easyesn [56]. To determine the optimal ESN hyperparameters a grid search is performed as described in [18] using the training and validation subsets of the data. This search consists of two stages: first, for each combination of the local states’ hyperparameters σ and as listed in Table 2 a grid search is performed to find the optimal five hyperparameters of the ESN resulting in 37 sets of optimal hyperparameters. To make these grid searches more feasible, they were performed just for a single input patch (area covered by the stencil, see Figure 5) in the spatial center of the training set and thus not using the full spatial data.
TABLE 2
| σ | σ | σ | σ | σ | σ | σ | σ | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 25 | 2 | 29 | 2 | 33 | 2 | 37 | 2 | 41 | 2 | 45 | 2 | 49 | 4 | 101 | 10 |
| 25 | 4 | 29 | 7 | 33 | 8 | 37 | 4 | 41 | 4 | 45 | 4 | 49 | 8 | 101 | 20 |
| 25 | 8 | 29 | 14 | 33 | 16 | 37 | 9 | 41 | 8 | 45 | 11 | 49 | 16 | 101 | 25 |
| 25 | 24 | 29 | 28 | 33 | 32 | 37 | 12 | 41 | 20 | 45 | 22 | 49 | 24 | 101 | 50 |
| 37 | 18 | 41 | 40 | 45 | 44 | 49 | 48 | ||||||||
| 37 | 36 |
The examined set of hyperparameters σ and for the local states.
TABLE 3
| CAE (MAE ± STD) | |||
|---|---|---|---|
| Case | Blurred data | Noisy data | Undersampled data |
| 1 | 0.01644 ± 0.00136 | 0.00794 ± 0.00096 | 0.00432 ± 0.00020 |
| 2 | 0.02076 ± 0.00170 | 0.00835 ± 0.00097 | 0.00782 ± 0.00053 |
| 3 | 0.02667 ± 0.00227 | 0.00856 ± 0.00104 | 0.01613 ± 0.00119 |
| 4 | 0.03450 ± 0.00318 | 0.00900 ± 0.00096 | 0.04727 ± 0.00393 |
| 5 | 0.04532 ± 0.00407 | 0.00919 ± 0.00099 | 0.12190 ± 0.01061 |
| 6 | 0.06137 ± 0.00585 | 0.00961 ± 0.00110 | 0.27821 ± 0.02823 |
| 7 | 0.08913 ± 0.00864 | 0.01210 ± 0.00103 | 0.42401 ± 0.02800 |
| 8 | 0.14018 ± 0.01261 | 0.01156 ± 0.00120 | - |
| 9 | 0.24689 ± 0.01898 | 0.01873 ± 0.00136 | - |
| 10 | 0.37214 ± 0.02408 | - | - |
| ESN (MAE ± STD) | |||
| Case | Blurred data | Noisy data | Undersampled data |
| 1 | 0.05220 ± 0.00347 | 0.06193 ± 0.00264 | 0.00362 ± 0.00031 |
| 2 | 0.05910 ± 0.00365 | 0.07193 ± 0.00288 | 0.01682 ± 0.00110 |
| 3 | 0.06245 ± 0.00394 | 0.08070 ± 0.00299 | 0.03516 ± 0.00242 |
| 4 | 0.07318 ± 0.00469 | 0.09052 ± 0.00312 | 0.08491 ± 0.00561 |
| 5 | 0.08476 ± 0.00536 | 0.09344 ± 0.00341 | 0.20439 ± 0.01105 |
| 6 | 0.09959 ± 0.00644 | 0.11136 ± 0.00370 | n.A. |
| 7 | 0.12325 ± 0.00813 | 0.11889 ± 0.00391 | n.A. |
| 8 | 0.18129 ± 0.01323 | 0.14548 ± 0.00596 | - |
| 9 | 0.27925 ± 0.01628 | 0.18259 ± 0.00720 | - |
| 10 | 0.39927 ± 0.01717 | - | - |
Comparison of the MAE obtained when applying the CAE method and the ESN method to the test data set.
In the second stage, for each of the 37 sets of optimal hyperparameters determined before (for each combination of σ and ), an ESN is trained on a larger subset of the training data and not just on a single patch. Ideally, this step should be performed on the entire spatial domain of the training set, however, as we did not notice significant differences in the results when the ESNs were trained on a spatial subset of size to speed up the training process. Following the same methodology as in [18], for each pixel from this spatial subset a single ESN is trained and then the obtained output matrices of these ESNs are averaged over all pixels. Compared to the procedure used in [18], the handling of boundary values has been changed. As for boundary pixels fewer adjacent pixels exist than for those inside, the creation of local states is obstructed, and boundary pixels require special treatment. In our previous work [18] individual ESNs have been trained for the boundary pixels using local states of lower dimensionality. In the following we use an alternative approach based on padding the boundary pixels by mirroring their values (motivated by the no-flux boundary conditions used). In this way, local states can be formally defined for boundary pixels in the same way as for inner pixels.
Next, the different optimal ESNs obtained for different stencils were evaluated by comparing their performance on the validation subset. In this way optimal values for σ and were selected by choosing the combination with the lowest difference between the prediction and ground truth on the validation set. This process yields an ESN whose hyperparameters and weights are optimized to yield minimal error. Finally, without training the network again on the entire training set, the optimal ESN found before is used to perform the prediction on the entire test set. As a pre-processing step, both the input and target data of the training, validation and test set are rescaled with min-max scaling, where the minimal and maximal value are determined over all pixels of the training set.
The CAE was trained using the ADAM optimizer [57], implemented with TensorFlow [58] in version 2.3, with early stopping when the validation loss has not improved at least by for 20 epochs. The learning rate was reduced by a factor of 0.2 when the loss metric stopped improving at least by for ten epochs. Dropout was set to be 0.05 in all cases. As loss function the mean absolute error (MAE) was chosen:where N is the number of elements in the data set, the network output, desired ground-truth and stands for the absolute values.
3.1.1 Noisy Data
Figure 8 shows snapshots of the noisy input data, the corresponding ground truth, the outputs provided by the CAE and the ESN, respectively, and the absolute values of their prediction errors with respect to the ground truth.
FIGURE 8
The evolution of the loss function during the training epochs is shown in Figure 9A. In all cases the error decreases and the training converges, but the duration of the training depends on the complexity of the case.
FIGURE 9
Figures 9B shows a comparison of the performance of the CAE and the ESN for noisy data with nine different noise levels . While the mean absolute error of the CAE remains below 0.02, the reconstruction error of the ESN increases from 0.06 for to 0.18 for , the associated ESN hyperparameter can be found in Table 4.
TABLE 4
| ESN hyperparameters for the case of noisy data | |||||
|---|---|---|---|---|---|
| Case | N | ρ | α | ε | L2 regularisation |
| 1 | 500 | 1.25 | 0.5 | 0.05 | 10 |
| 2 | 500 | 0.50 | 0.2 | 0.05 | 10 |
| 3 | 250 | 1.00 | 0.2 | 0.2 | 10 |
| 4 | 250 | 0.05 | 0.2 | 0.2 | 10 |
| 5 | 250 | 3.00 | 0.2 | 0.1 | 10 |
| 6 | 250 | 1.25 | 0.2 | 0.1 | 10 |
| 7 | 250 | 3.00 | 0.2 | 0.05 | 10 |
| 8 | 500 | 1.25 | 0.05 | 0.2 | 10 |
| 9 | 500 | 3.00 | 0.05 | 0.05 | 10 |
Selected ESN hyperparameters for the case of noisy data.
3.1.2 Blurred Data
To evaluate the performance of CAE and ESN for recovering full resolution (ground truth) data from blurred observations we consider nine cases where the radius m of Fourier low-pass filtering ranges from to (in steps of 2). Figure 10 shows snapshots of reconstructions of the u-variable of the BOCF model using CAE and ESN with filter parameters (A-F), (G-L), and (M-R). Similar to Figure 8 the errors are largest at fronts of the excitation waves, but in contrast to noisy images the performances of CAE and ESN differ not much for blurred data. This observation is also confirmed by a systematic comparison of the mean absolute errors of both methods for different manhatten distances m given in Figure 11A. The errors decrease with m because the larger m the less blurred are the input data of the CAE or ESN (for hyperparameter see Table 5). Figure 11B shows the evolution of the loss function during training of the CAE.
TABLE 5
| ESN hyperparameters for the case of noisy data | |||||
|---|---|---|---|---|---|
| Case | N | ρ | α | ε | L2 regularisation |
| 1 | 250 | 0.30 | 0.01 | 0.1 | 10 |
| 2 | 250 | 1.25 | 0.05 | 0.05 | 10 |
| 3 | 500 | 2.00 | 0.05 | 0.1 | 10 |
| 4 | 500 | 0.75 | 0.2 | 0.2 | 5 |
| 5 | 250 | 1.50 | 0.2 | 0.2 | 10 |
| 6 | 250 | 1.50 | 0.2 | 0.2 | 10 |
| 7 | 250 | 1.25 | 0.2 | 0.2 | 10 |
| 8 | 500 | 0.90 | 0.2 | 0.2 | 5 |
| 9 | 250 | 1.50 | 0.5 | 0.2 | 10 |
| 10 | 500 | 0.75 | 0.7 | 0.2 | 10 |
Selected ESN hyperparameters for the case of blurred data.
FIGURE 10
FIGURE 11
3.1.3 Undersampled Data
Figure 12 shows examples of data reconstructed from undersampled data. In total we considered seven cases of undersampling by pixels, where i ranges from to . For input images have a resolution of and for , . In all cases the desired output (ground truth) has a size of pixels. The used hyperparameter for the ESN can be found in Table 6). Figure 13B shows the evolution of the corresponding loss function.
FIGURE 12
TABLE 6
| ESN hyperparameters for the case of noisy data | |||||
|---|---|---|---|---|---|
| Case | N | ρ | α | ε | L2 regularisation |
| 1 | 250 | 0.30 | 0.01 | 0.1 | 5 |
| 2 | 500 | 0.05 | 0.05 | 0.2 | 0.5 |
| 3 | 250 | 1.50 | 0.05 | 0.2 | 10 |
| 4 | 250 | 2.00 | 0.5 | 0.2 | 10 |
| 5 | 500 | 0.30 | 0.05 | 0.1 | 5 |
| 6 | 500 | 0.75 | 0.2 | 0.05 | 10 |
| 7 | 250 | 1.50 | 0.05 | 0.1 | 10 |
Selected ESN hyperparameters for the case of undersampled data.
FIGURE 13
3.2 Predicting Electrical Excitation From Mechanical Contraction
Echo state networks as well as convolutional autoencoders have been trained with time series generated by the electromechanical model introduced in Section 2.2 to predict the membrane voltage Eq. (4) from the local contraction given in Eq. (13). The sampling time of all time series equalled . Since for periodically rotating spiral waves this cross-prediction task is quite straightforward we focus here on the much more demanding case of an example exhibiting spatio-temporal chaos (corresponding to atrial or ventricular fibrillation). Figure 14 shows at three instants of time snapshots of the observed contraction (first column), the ground truth of the voltage u (second column), the prediction of the CAE and its absolute error (third and fourth column) and the prediction of the ESN and the corresponding absolute error (columns five and six, respectively). Both, the ESN and the CAE were trained and tested with the same spatio-temporal time series (lengths of training, validation and test sets are 15,000, 2000 and 2000 samples, respectively). Hyperparameters of the ESN are and . To make the ESN more robust normally distributed noise with zero mean and a variance of was added to the arguments of the activation function.
FIGURE 14
As illustrated in Figure 14 both networks can solve the inverse problem and reconstruct the electrical potential field u from Eq. (4). However, the reconstruction of the CAE is more precise, which is particularly noticeable at the edges of the reconstructed electrical potential field. Considering the median of the MAE over the entire test data the ESN approach achieves an error of while the median error of the CAE equals .
4 Conclusion
Using synthetic data generated with conceptual models describing complex cardiac dynamics we have demonstrated possible applications of machine learning methods to complete and enhance experimental observations. It was shown that echo state networks as well as convolution autoencoders provide promising results, where the latter turned out to be the method of choice in terms of more faithful reconstructions. At this point, however, we would like to stress, that we didn’t try to fully optimize the algorithms employed. One could, for example, increase the size of the ESNs used or extend and refine the grid search of hyperparameters. Also with the CAE several options exist to improve the performance even more. Instead of the MAE in the loss function one could use an adaptive robust loss function [59] or the Jensen-Shannon divergence [19]. The weights of the CAE could be optimized with a stochastic gradient descend approach instead of the ADAM algorithm [57]. But we expect such modifications would show only minor improvements (if at all) [60, 61].
A comparison of the computing times with ESNs and CAEs is unfortunately not immediately possible. For the CAE the training time depends on the convergence, as illustrated in Figures 9A, 11B and 13B. In contrast to this, for ESNs training times depend strongly on the search for optical hyperparameters, especially the size of the reservoir, and this search is strongly dependent on the search space size and the number of parameters. For the task where electrical excitation is predicted from mechanical contraction our computations took 3,382 s on two Intel Xeon CPU E5-4620. While the CAE simulations have been run on GPUs the training and application of the ESN was on CPUs which makes a direct comparison difficult. Furthermore, the runtime of the programmes used is highly dependent on the libraries used and how well they have been adapted to special system architectures. In general we would estimate that in this work the effort to train and search hyperparameters for an ESN was a bit less demanding, in the sense of computational resources, compared to the training of the CAE. However, we would not consider the difference big enough to be an advantage for the ESN approach. Once trained both approaches need comparable execution times when applied to new data and executed on a CPU. In future work, using more realistic numerical simulations (and experimental data) such an optimization should be performed to achieve the best possible result for the intended medical application. Since here we used only data from conceptual models we refrained from fully optimizing the machine learning methods applied. The fact that already a straight-forward application of known algorithms and architectures provided very good results for the considered reconstruction tasks is very promising and encourages to address in future work extended tasks (including other variables, like calcium concentration, mechanical stress and strain, etc.) and reconstruction tasks with more realistic synthetic data (from 3D models, for example) combined with experimental measurements.
Statements
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation. The source code will be available at https://gitlab.gwdg.de/sherzog3/reconstructing-cardiac-excitation-waves.git.
Author contributions
SH performed the CAE simulations and RZ and JA did the ESN modelling. All authors designed the study, analyzed the results, and wrote the manuscript.
Funding
SH acknowledges funding by the International Max Planck Research Schools of Physics of Biological and Complex Systems. UP and SL acknowledge funding by the Max Planck Society, the German Center for Cardiovascular Research (DZHK) partner site Goettingen, and the SFB 1002 Modulatory Units in Heart Failure (Project C03).
Acknowledgments
SH and UP thank Florentin Wörgötter and Thomas Lilienkamp for continuous support and inspiring scientific discussions.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
References
1.
GrayRAPertsovAMJalifeJ. Spatial and temporal organization during cardiac fibrillation. Nature (1998). 392:75–8. 10.1038/32164
2.
DavidenkoJMPertsovAVSalomonszRBaxterWJalifeJ. Stationary and drifting spiral waves of excitation in isolated cardiac muscle. Nature (1992). 355:349–51. 10.1038/355349a0
3.
WitkowskiFXLeonLJPenkoskePAGilesWRSpanoMLDittoWLet alSpatiotemporal evolution of ventricular fibrillation. Nature (1998). 392:72–82. 10.1038/32170
4.
EfimovIRNikolskiVPSalamaG. Optical imaging of the heart. Circ Res (2004). 95:21–33. 10.1161/01.RES.0000130529.18016.35
5.
RosenzweigSScholandNHolmeHCMUeckerM. Cardiac and respiratory self-gating in radial mri using an adapted singular spectrum analysis (ssa-fary). IEEE Trans Med Imag (2020). 39:3029–3041.
6.
OtaniNLutherSSinghRGilmourR. Transmural ultrasound-based visualization of patterns of action potential wave propagation in cardiac tissue. Ann Biomed Eng (2010a). 38:3112–23.
7.
OtaniNLutherSSinghRGilmourR. Methods and systems for functional imaging of cardiac tissueInternational Patent. London: World Intellectual Property Organization (2010b).
8.
ProvostJLeeWNFujikuraKKonofagouEE. Imaging the electromechanical activity of the heart in vivo. Proc Natl Acad Sci USA (2011). 108:8565–70. 10.1073/pnas.1011688108
9.
BersDM. Cardiac excitation-contraction coupling. Nature (2002). 415:198–205. 10.1038/415198a
10.
BourgeoisEBBachtelADHuangJWalcottGPRogersJM. Simultaneous optical mapping of transmembrane potential and wall motion in isolated, perfused whole hearts. J Biomed Optic (2011). 16:096020. 10.1117/1.3630115
11.
ZhangHIijimaKHuangJWalcottGPRogersJM. Optical mapping of membrane potential and Epicardial deformation in beating hearts. Biophys J (2016). 111:438–51. 10.1016/j.bpj.2016.03.043
12.
ChristophJSchröder-ScheteligJLutherS. Electromechanical optical mapping. Prog Biophys Mol Biol (2017). 130:150–69. 10.1016/j.pbiomolbio.2017.09.015
13.
ChristophJChebbokMRichterCSchroeder-ScheteligJSteinSUzelacIet alElectromechanical vortex filaments during cardiac fibrillation. Nature (2018). 555:667–72. 10.1038/nature26001
14.
BergSLutherSParlitzU. Synchronization based system identification of an extended excitable system. Chaos: An Interdisciplinary Journal of Nonlinear Science (2011). 21:033104. 10.1063/1.3613921
15.
LebertJChristophJ. Synchronization-based reconstruction of electromechanical wave dynamics in elastic excitable media. Chaos: An Interdisciplinary Journal of Nonlinear Science (2019). 29:093117. 10.1063/1.5101041
16.
HoffmanMJLaVigneNSScorseSTFentonFHCherryEM. Reconstructing three-dimensional reentrant cardiac electrical wave dynamics using data assimilation. Chaos Interdiscipl J Nonlinear Sci (2016). 26:013107. 10.1063/1.4940238
17.
HoffmanMJCherryEM. Sensitivity of a data-assimilation system for reconstructing three-dimensional cardiac electrical dynamics. Phil Trans Math Phys Eng Sci (2020). 378:20190388. 10.1098/rsta.2019.0388
18.
ZimmermannRSParlitzU. Observing spatio-temporal dynamics of excitable media using reservoir computing. Chaos Interdiscipl J Nonlinear Sci (2018). 28:043118. 10.1063/1.5022276
19.
HerzogSWörgötterFParlitzU. Data-driven modeling and prediction of complex spatio-temporal dynamics in excitable media. Front Appl Math Stat (2018). 4:60. 10.3389/fams.2018.00060
20.
HerzogSWörgötterFParlitzU. Convolutional autoencoder and conditional random fields hybrid for predicting spatial-temporal chaos. Chaos (2019). 29:123116. 10.1063/1.5124926
21.
ChristophJLebertJ. Inverse mechano-electrical reconstruction of cardiac excitation wave patterns from mechanical deformation using deep learning. Chaos (2020). 30:123134.10.1063/5.0023751
22.
Bueno-OrovioACherryEMFentonFH. Minimal model for human ventricular action potentials in tissue. J Theor Biol (2008). 253:544–60. 10.1016/j.jtbi.2008.03.029
23.
ClaytonRBernusOCherryEDierckxHFentonFMirabellaLet alMathematical and modelling foundations, models of cardiac tissue electrophysiology: progress, challenges and open questionsProg Biophys Mol Biol (2011). 104:22–48. 10.1016/j.pbiomolbio.2010.05.008
24.
Ten TusscherKNobleDNoblePPanfilovAV. A model for human ventricular tissue. Am J Physiol Heart Circ Physiol (2004). 286:H1573–H1589.
25.
StrainMCGreensideHS. Size-dependent transition to high-dimensional chaotic dynamics in a two-dimensional excitable medium. Phys Rev Lett (1998). 80:2306–9. 10.1103/PhysRevLett.80.2306
26.
LilienkampTChristophJParlitzU. Features of chaotic transients in excitable media governed by spiral and scroll waves. Phys Rev Lett (2017). 119:054101.
27.
AlievRRPanfilovAV. A simple two-variable model of cardiac excitation. Chaos Solit Fractals (1996). 7:293–301. 10.1016/0960-0779(95)00089-5
28.
NashMPPanfilovAV. Electromechanical model of excitable tissue to study reentrant cardiac arrhythmias. Prog Biophys Mol Biol (2004). 85:501–22. 10.1016/j.pbiomolbio.2004.01.016
29.
GöktepeSKuhlE. Electromechanics of the heart: a unified approach to the strongly coupled excitation–contraction problem. Comput Mech (2009). 45:227–43. 10.1007/s00466-009-0434-z
30.
ErikssonTSEPrasslAPlankGHolzapfelG. Influence of myocardial fiber/sheet orientations on left ventricular mechanical contraction. Math Mech Solid (2013). 18:592–606. 10.1177/1081286513485779
31.
BourguignonDCaniMP. Controlling anisotropy in mass-spring systems. Eurographics. Vienna: Springer (2000). p. 113–23. 10.1007/978-3-7091-6344-3_9
32.
JaegerHGMD Report. The ‘echo state’ approach to analysing and training recurrent neural networks—with an erratum note. Kaiserslauten: German National Research Institute for Computer Science (2001). p. 43.
33.
ChengZSunHTakeuchiMKattoJ. Deep convolutional autoencoder-based lossy image compressionPicture Coding Symposium, PCS 2018—Proceedings. London: Institute of Electrical and Electronics Engineers Inc. (2018). p. 253–7. 10.1109/PCS.2018.8456308
34.
LukoševičiusM. A practical guide to applying echo state networks. Lecture notes in computer science. Berlin: Springer (2012). p. 659–86. 10.1007/978-3-642-35289-8_36
35.
LukoševičiusMJaegerH. Reservoir computing approaches to recurrent neural network training. Comput Sci Rev (2009). 3:127–49. 10.1016/j.cosrev.2009.03.005
36.
YildizIBJaegerHKiebelSJ. Re-visiting the echo state property. Neural Netw (2012). 35:1–9. 10.1016/j.neunet.2012.07.005
37.
JaegerHHaasH. Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless communication. Science (2004). 304:78–80. 10.1126/science.1091277
38.
LuZPathakJHuntBGirvanMBrockettROttE. Reservoir observers: model-free inference of unmeasured variables in chaotic systems. Chaos Interdiscipl J Nonlinear Sci (2017). 27:041102. 10.1063/1.4979665
39.
CarrollTLPecoraLM. Network structure effects in reservoir computers. Chaos Interdiscipl J Nonlinear Sci (2019). 29:083130. 10.1063/1.5097686
40.
GriffithAPomeranceAGauthierDJ. Forecasting chaotic systems with very low connectivity reservoir computers. Chaos Interdiscipl J Nonlinear Sci (2019). 29:123108. 10.1063/1.5120710
41.
ThiedeLAParlitzU. Gradient based hyperparameter optimization in echo state networks. Neural Network (2019). 115:23–9. 10.1016/j.neunet.2019.02.001
42.
HaluszczynskiAAumeierJHerteuxJRthC. Reducing network size and improving prediction stability of reservoir computing. Chaos Interdiscipl J Nonlinear Sci (2020). 30:063136. 10.1063/5.0006869
43.
CarrollTL. Dimension of reservoir computers. Chaos Interdiscipl J Nonlinear Sci (2020). 30:013102. 10.1063/1.5128898
44.
FanHJiangJZhangCWangXLaiYC. Long-term prediction of chaotic systems with machine learning. Phys Rev Res (2020). 2:012080. 10.1103/PhysRevResearch.2.012080
45.
ParlitzUHornsteinA. Dynamical prediction of chaotic time series. Chaos Complex Lett (2005). 1:135–44.
46.
LuZHuntBROttE. Attractor reconstruction by machine learning. Chaos Interdiscipl J Nonlinear Sci (2018). 28:061104. 10.1063/1.5039508
47.
PathakJHuntBGirvanMLuZOttE. Model-free prediction of large spatiotemporally chaotic systems from data: a reservoir computing approach. Phys Rev Lett (2018). 120. 10.1103/physrevlett.120.024102
48.
ParlitzUMerkwirthC. Prediction of spatiotemporal time series based on reconstructed local states. Phys Rev Lett (2000). 84:1890–3. 10.1103/physrevlett.84.1890
49.
LeCunYBoserBEDenkerJSHendersonDHowardREHubbardWEet alHandwritten digit recognition with a back-propagation network. In: TouretzkyDS, editor. Advances in neural information processing systems, 2Burlington, MA: Morgan-Kaufmann (1990). p. 396–404.
50.
IoffeSSzegedyC. Batch normalization: accelerating deep network training by reducing internal covariate shift. Proc Int Conf Mach Learn (2015). 37:448456. 10.1609/aaai.v33i01.33011682
51.
MaasALHannunAYNgAY. “Rectifier nonlinearities improve neural network acoustic models Speech and Language Processing.” In ICML workshop on deep learning for audio (2013). p. 3.
52.
HahnloserRHSarpeshkarRMahowaldMADouglasRJSeungHS. Digital selection and analogue amplification coexist in a cortex-inspired silicon circuit. Nature (2000). 405:947.
53.
SrivastavaNHintonGKrizhevskyASutskeverISalakhutdinovR. Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res (2014). 15:1929–58. 10.1109/iwcmc.2019.8766500
54.
DumoulinVVisinF. A guide to convolution arithmetic for deep learning (2016). arXiv e-printsarXiv:1603.07285
55.
BäuerleARopinskiT. Net2vis: transforming deep convolutional networks into publication-ready visualizations (2019). arXiv preprint arXiv:1902.04394.
56.
ZimmermannRThiedeL. easyesn (2020). Available at: https://github.com/kalekiu/easyesn.
57.
KingmaDPBaJ. Adam: a method for stochastic optimization (2014). arXiv e-printsarXiv:1412.6980.
58.
AbadiMAgarwalABarhamPBrevdoEChenZCitroCet alTensorFlow: large-scale machine learning on heterogeneous systems (2015). Software available from tensorflow.org.
59.
BarronJT. A general and adaptive robust loss function. Berlin: Springer (2019).
60.
RosserJB. Nine-point difference solutions for poisson's equation. Comput Math Appl (1975). 1:351–60. 10.1016/0898-1221(75)90035-8
61.
SchererPOJ. Computational physics: simulation of classical and quantum systems. Berlin: Springer (2010). p. 147.
62.
ChughSSHavmoellerRNarayananKSinghDRienstraMBenjaminEJet alWorldwide epidemiology of atrial fibrillation: a Global Burden of Disease 2010 Study. Circulation (2014)129:837–847. 10.1161/CIRCULATIONAHA.113.005119
63.
WolfPADawberTRThomasHEKannelWB. Epidemiologic assessment of chronic atrial fibrillation and risk of stroke: the Framingham study. Neurology (1978)28:973–977.
64.
ManolioTAKronmalRABurkeGLO'LearyDHPriceTR. Short-term predictors of incident stroke in older adults. The Cardiovascular Health Study. Stroke (1996)27:1479–1486.
Appendix 1: Solution of the electro-mechanical dynamics
The differential equations from the extended Aliev-Panfilov model in Eqs. A1–A3 have been integrated using the forward Euler method , in which y is a place holder for the model variables u, v and ; represents the right-hand side of the respective equation. The diffusion tensor in Equation (4) was set to a scalar constant D and the diffusion term was approximated with a nine-point stencil [60]where i, j are the indices of the grid points and h denotes the spacing constant between the cells.
For the excitation variable u in the electrical part of the simulation, no-flux boundary conditions have been used which were imposed by setting the two outermost cells to the same value, i.e., . In the mechanical part of the simulation, numerical calculations were carried out according to the following scheme for each time step: (1) update of the position of the centre of mass , (2) calculation of all four points of attachment for each cell, (3) computing of forces from structural and axial springs for each particle , (4) update of the positions of all particles using the Verlet method and (5) determine change of area for each cell. Here the Verlet method refers to the standard Verlet algorithm which is given as [61]with the total force acting on the particle. Because the total force includes the damping term, it is convenient to rewrite Eq. A2 with to
A padding layer of ten electrically inactive cells was implemented outside the electrical grid to account for boundary effects in the mechanical network. In addition, the active stress variable from Eq. (6) of the last row of cells just at the edge of the simulation grid was mirrored to the two padding layers just outside the simulation grid which proved to dramatically reduce mechanical boundary effects. This is likely due to the fact that a proper contraction of an electrically active cell is not guaranteed if one of its sides is connected to an inactive cell. Lastly, the outermost padding cell’s positions were fixed to prevent the grid as a whole from moving away from its original position.
To improve numerical accuracy for each time step of the electrical equations (Euler method) five time steps of size were computed for the mechanical system (3). All computations have been performed on a spatial grid of 100 100 elements. The parameter values of the dynamical equations used are given in Table A1.
TABLE A1
| 0.1 | a | 0.05 | b | 0.05 | 0.2 | 0.3 | k | 8 | 0.002 | ||||
| D | 0.22 | 0.08 | 3 | 13 | 23 | 2 | 9 | ||||||
| 23 | 10 | 0.2 | ν | 6.86 | 1 | 0.1 | 30 |
Parameters of the electro-mechanical model.
Summary
Keywords
reservoir computing, convolutional autoencoder, image enhancement, cross-prediction, cardiac arrhythmias, excitable media, electro-mechanical coupling, cardiac imaging
Citation
Herzog S, Zimmermann RS, Abele J, Luther S and Parlitz U (2021) Reconstructing Complex Cardiac Excitation Waves From Incomplete Data Using Echo State Networks and Convolutional Autoencoders. Front. Appl. Math. Stat. 6:616584. doi: 10.3389/fams.2020.616584
Received
12 October 2020
Accepted
07 December 2020
Published
18 March 2021
Volume
6 - 2020
Edited by
Axel Hutt, Inria Nancy - Grand-Est Research Centre, France
Reviewed by
Petia D. Koprinkova-Hristova, Institute of Information and Communication Technologies (BAS), Bulgaria
Meysam Hashemi, INSERM U110 6 Institut de Neurosciences des Systèmes, France
Updates
Copyright
© 2021 Herzog, Zimmermann, Abele, Luther and Parlitz.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Ulrich Parlitz, ulrich.parlitz@ds.mpg.de
This article was submitted to Dynamical Systems, a section of the journal Frontiers in Applied Mathematics and Statistics
Disclaimer
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.