Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Physiol., 18 August 2025

Sec. Computational Physiology and Medicine

Volume 16 - 2025 | https://doi.org/10.3389/fphys.2025.1629121

Echo-ODE: A dynamics modeling network with neural ODE for temporally consistent segmentation of video echocardiograms

Wenliang LuWenliang Lu1Yuan Wang&#x;Yuan Wang 2Wenli Dai&#x;Wenli Dai 1Yingnan Wu&#x;Yingnan Wu 3Hao Xu,
Hao Xu4,5*Dexing Kong
Dexing Kong1*
  • 1School of Mathematical Sciences, Zhejiang University, Hangzhou, Zhejiang, China
  • 2Department of Mathematics, Zhejiang Sci-Tech University, Hangzhou, Zhejiang, China
  • 3Department of Ultrasound Medicine, Litao Sun Cancer Center, Zhejiang Provincial People’s Hospital (Affiliated People’s Hospital), Hangzhou Medical College, Hangzhou, Zhejiang, China
  • 4College of Mathematical Medicine, Zhejiang Normal University, Jinhua, Zhejiang, China
  • 5Cardiovascular Research Group, Puyang Institute of Big Data and Artificial Intelligence, Puyang, Henan, China

Introduction: Segmentation of echocardiograms plays a crucial role in clinical diagnosis. Beyond accuracy, a major challenge of video echocardiogram analysis is the temporal consistency of consecutive frames. Stable and consistent segmentation of cardiac structures is essential for a reliable fully automatic echocardiogram interpretation.

Methods: We propose a novel framework Echo-ODE, where the heart is regarded as a dynamical system, and we model the representation of dynamics by neural ordinary differential equations. Echo-ODE learns the spatio-temporal relationships of the input video and output continuous and consistent predictions.

Results: Experiments conducted on the Echo-Dynamic, the CAMUS and our private dataset demonstrate that Echo-ODE achieves comparable accuracy but significantly better temporal stability and consistency in video segmentation than previous mainstream CNN models. More accurate phase detection and robustness to arrhythmia also underscore the superiority of our proposed model.

Discussion: Echo-ODE addresses the critical need for temporal coherence in clinical video analysis. This framework establishes a versatile backbone extendable beyond segmentation tasks. Its ability to model cardiac dynamics demonstrates great potential for enabling reliable, fully automated video echocardiogram interpretation. The code is publicly available at https://github.com/luwenlianglu/EchoODE.

1 Introduction

Accurate assessment of cardiac structure and function is crucial to clinical diagnosis and guide optimal treatment options for patients (Alsharqi et al., 2018). Echocardiography is one of the most widely used imaging modalities that allows real-time imaging and can detect various abnormalities. A key aspect of echocardiography is the assessment and quantification of cardiac chamber size and function, particularly the left ventricle (LV). Quantification of left ventricular ejection fraction (LVEF), for example, is the central measure of LV systolic function (Loehr et al., 2008; Huang et al., 2017), and it is calculated by the fraction of LV volume of the end-systolic (ES) and end-diastolic (ED). Currently, sonographers typically determine the cardiac phase of an echocardiogram visually. They track ED frames and ES frames through consecutive cardiac cycles and then manually delineate the chamber boundary to calculate LVEF and other functional metrics. This process is time-consuming, tedious and prone to poor repeatability (Koh et al., 2017). Furthermore, LVEF assessment is recommended to average the results of several consecutive cardiac cycles if variation is identified. However, in practice, it is often evaluated from tracings of just one representative beat (Lang et al., 2015), leading to high variance and limited precision due to inter-observer variation (Malm et al., 2004; Farsalinos et al., 2015). While automatic segmentation of video echocardiograms can solve the problem, temporally inaccurate or inconsistent segmentation can lead to incorrect phase detection, resulting in unreliable LVEF calculations.

Early attempts to algorithmically assess cardiac function applied deep learning models to manually curated still images on ED frames and ES frames (Zhang et al., 2018; Moradi et al., 2019; Moradi et al., 2019; Liu et al., 2021; Guo et al., 2021; Su et al., 2024; Akbari et al., 2024), or segmenting multiple planes in one model Hu et al. (2025). Some subsequent works extended these image-level models to echocardiogram videos, analyzing them frame-by-frame and plotting the curve of left ventricular volume over time. The peak and valley points of the curve were recognized as ED frames and ES frames (Ouyang et al., 2020; Duffy et al., 2022). Although these models allowed fully automated cardiac function assessment, they ignored the temporal dependencies between video frames. Later models sought to combine spatial and temporal information (Lane et al., 2021; Wu et al., 2022; Zeng et al., 2023; Ta et al., 2024; Hasan et al., 2025). These models processed a few consecutive frames as input and predicted the label for the last frame. Models in this category were still considered image-level models, as only a single frame prediction could be derived from each computation. The approaches mentioned above either applied image-level nets to each frame or used consecutive frames as input but only output a one-frame prediction, and none of them guaranteed temporal consistency. Therefore, it is necessary to promote a temporally stable and consistent segmentation algorithm for video echocardiograms.

Inspired by reference (Özçelik and Altan, 2023) to deal with non-linear dynamics, we sought a new framework capable of overcoming these challenges and specifically designed to handle such a unique system. Recall that a dynamical system consists of an abstract phase or state space, where the coordinates describe the system’s state at any given moment, and a dynamical rule that specifies the future evolution of the system based solely on the present values of its state variables. Intuitively, the heart clearly represents a biological dynamical system. The processes of diastole and systole over time constitute the state space. The heart beats regularly controled by the cardiac conduction system, allowing us to learn the dynamic rule and predict future states. This dynamic property still holds when applied to video echocardiograms, which represent a sectional view or subspace of the entire heart space. Mathematically, a dynamical system is described by the initial value problem of a differential equation, which inspired us to model the echocardiogram using neural ordinary differential equations (NODE). Notably, many differential equation solvers include a scheme known as dense output, which generates high-order interpolations by reusing internal computations from the steps without requiring additional computational resources or time (Shampine and Jay, 2015). The advantages of modeling echocardiograms with differential equations are as follows: (1) The heart is treated as a dynamical system, making the model more interpretable; (2) The input series does not need to have a constant time duration, unlike the requirement in RNN models; (3) The outputs of NODE are dense and continuous in the time dimension, allowing us to test our model on real videos with non-adjacent input frames. Intermittent sampling is possible, and internal predictions can be generated through dense outputs, enhancing computational efficiency; (4) The continuous outputs of NODE make it feasible to perform some video generation tasks, such as video extrapolation and interpolation on sparsely sampled echocardiograms.

2 Materials and methods

2.1 Neural ordinary differential equations

NODEs (Chen et al., 2018) represent a family of parameterized algorithms designed to model the evolution over time of a system with state ξ(t) at an arbitrary time t. These systems are governed by continuous-time dynamics that satisfy a Cauchy (or initial value) problem:

ξtt=fξt,t,ξt0=ξ0.

By approximating the differential with an estimator fθf parameterized by θ, such as a neural network, these methods enable learning of such dynamics (or trajectories) from relevant data. With this formalization, the state ξ(t) of the system is defined at all times, and can be computed at any desired time using a numerical ODESolver

ξt0,ξt1,,ξtn=ODESolverfθ,ξ0,t0,t1,,tn.

For any single arbitrary time value ti, a call to ODESolver computes a numerical approximation of the integral of dynamics from the initial time value t0 to ti.

ξti=ODESolverfθ,ξ0,t0,ti=ξ0+t0tifθξs,sds.

A plethora of algorithms for numerical integration of differential equations can be found in the existing literature. The optimization of neural ODEs is carried out within the framework of adjoint sensitivity (Pontryagin, 1985).

Inspired by the application of NODEs to continuous time series modeling, the latent ODE (Rubanova et al., 2019) equipped with an ODE-RNN was proposed to handle irregularly sampled time series data. Ma et al. (2022) introduced CortexODE, which leveraged NODEs to learn a diffeomorphic flow for the reconstruction of the cortical surface, ensuring that the reconstruction preserved the topology. ODE2VAE (Yildiz et al., 2019) aimed to decompose latent representations into position and momentum to generate low-resolution image sequences. Park et al. (2021) proposed Vid-ODE, a network for continuous-time video generation.

2.2 Notation and overview

2.2.1 Notation

We denote Xτ{Xt1,Xt2,,Xtn} as a sequence of input video frames of length n, where each XtiRh×w×c is a 2D image of size h×w with c channels sampled at either regularly or irregularly spaced timesteps τ{t1,t2,,tn}, with 0<t1<t2<<tn. Given an input video Xτ and a specific task T for another set of time steps S{s1,s2,,sm}, our goal is to generate the predictions TS{Ts1,Ts2,,Tsm} at these timesteps. S needs not to be the same as τ and can be any countable subset of the interval [0,tn]. The origin is defined as t0=s0=0.

2.2.2 Overview

An overview of our proposed Echo-ODE framework is shown in Figure 1. Echo-ODE adopts an Encoder-Decoder structure. First, a convolutional encoder is applied to reduce the dimension and capture the spatial features of all frames in the input clip with shared parameters, it is followed by a GRU-ODE (Latent ODE) block to capture the temporal dependencies and learn the overall infomation and initial state of the video (top left of Figure 1, details described in Section 2.3). These outputs are then fed into an ODESolver to obtain continuous hidden states at arbitrary time steps. These hidden states are used to generate the desired predictions for a certain task by a target decoder (bottom right of Figure 1). To enable the model to learn the essential representations of the video, this paper adds a self-supervised video reconstruction block (top right of Figure 1), which is also a shared parameter convolutional network. A skip connection from the reconstruction decoder to the target decoder enhances the target decoder’s ability to capture high-resolutional semantic features (details described in Section 2.4).

Figure 1
Diagram illustrating Echo-ODE for processing echocardiogram video. From left to right, ultrasound images are encoded by a convolutional encoder with shared parameters and processed by a sequence of GRU-ODE blocks. The outputs are fed into an ODESolver, followed by a reconstruction decoder, a target decoder, and skip-connections between the two decoders to generate the desired continuous predictions. The lower section details the derivative function, involving concatenation and 3x3 convolutional operations.

Figure 1. Overview of our proposed Echo-ODE. A convolutional encoder embedded the input video clip into the hidden states. It was followed by the latent ODE, ODESolver and convolutional decoders to generated the target prediction. Xtn: input images. X̂sn: reconstructed image. Tsn: target output, generated by adjusting the decoder head according to different tasks.

2.3 Encoder

Given the input {Xti}, a convolutional encoder E (top left of Figure 1) is applied to learn the spatial feature and embed the video frames into hidden states {ĥti}. The encoder consists of four pairs of convolutional layers with max-pooling between adjacent convolutional pairs, resulting in four different resolutions. The number of 3×3 convolutional kernels is (32, 32), (64, 64), (128, 128), (256, 256) in each pair. The encoder shares the same parameters across different frames of the clip. Extracted features can be expressed as:

ĥti=EXti.

It is followed by a latent ODE (Rubanova et al., 2019) to capture the temporal relationships and learn the overall dynamic information D. An initial state hs0 is derived from D by a convolutional layer. The latent ODE is a combination of neural ODE (Chen et al., 2018) and ConvGRU (Ballas et al., 2015). While NODE can successfully model temporal dynamics at irregular time steps, it cannot directly process media input. On the other hand, RNN models struggle with irregular time input. To address this issue, the latent ODE is designed to capture the temporal dependencies with irregular timestep media input.

To obtain a better initial state hs0 that will be as one of the inputs of ODESolver, we apply the latent ODE backward-in-time as suggested by the authors of Rubanova et al. (2019), i.e., from the last frame of the input clip to the first one. In our case, the latent ODE can be formulated as:

hti1=ODESolverf,hti,ti,ti1,hti1=ConvGRUCellhti1,EXti1,

where i=n,n1,,1. The ODE solver calculates the next hidden state hti1 by integration dhtidt, which is approximated by a neural network f. Then hti1 is updated with an encountered media input E(Xti1) at the current time using a ConvGRU cell to produce hti. The initial hidden state htn+1=htn+ϵ is set to zeros, with ϵ set to 0.01. The obtained hs0 serves as the initial value of the NODE decoder, and the overall dynamic information D is integrated into the neural derivative function.

2.4 Decoder

2.4.1 Baseline

The decoder uses NODE to generate a sequence of frames at the target timesteps {s1,s2,,sm} based on the latent representation of the initial state hs0 and the derivative function fθ(D,θ). The newly generated hidden state at any time si can be presented as:

hsi=hs0+s0sifθD,θds,i=1,2,,m.

This process can be solved by any ODESolver.

We observe that the derivative function fθ should be unique for each video clip because every heart follows its own dynamic pattern. To the best of our knowledge, models that apply NODE to time series up to now only learn the initial state, and the derivative function is fixed once the training process is complete. These models typically perform training and testing on the same time series to accomplish tasks such as interpolation and extrapolation. In other words, these models fine-tune the parameters for each sample individually, which is clearly not suitable for our case. The reason for these limitations is that their derivative functions are fixed and only receive the state variable as input. To address this limitation, we integrate the heart dynamic information learned from each clip into the derivative function, as shown in the bottom left of Figure 1. For any time t, the function concatenates the hidden state ht and the dynamic information D, which is then passed through three convolutional layers to compute the derivative. This derivative function dynamically adjusts to different video inputs through D, eliminating the need for video-specific training.

Note that most ODE solvers employ a scheme known as dense output to generate high-order interpolations by reusing the internal computations from the solver steps (Shampine and Jay, 2015). This interpolation scheme often requires no additional calculations, allowing the target time steps to be any subset of time, regardless of whether the time series is regular or irregular, sparse or dense. Due to the dense output of NODE, we can obtain intermediate predictions even when the input consists of sparsely sampled frames, thereby improving computational efficiency.

2.4.2 Dynamic representation learning

Theoretically, a target decoder (bottom right of Figure 1) can accomplish our task on its own, and it is provided as a baseline for our method in this work. However, video echocardiograms are typically annotated on only one or two frames in practice, making it difficult for our model to fully understand the dynamics of the heart by relying solely on sparsely annotated task labels.

To address this issue, we add a reconstruction block as a self-supervised learning strategy to capture the essential representation of the video (top right of Figure 1). The reconstruction decoder consists of a sequence of upsampling convolutional layers that reverse the encoder process and produce three-channel reconstructed images. Since the output time steps do not have to be the same as the input, this block functions as image reconstruction when the target time node coincides with one of the input time steps. Otherwise, it can be interpreted as video interpolation. Both video reconstruction and video interpolation encourage the network to gain a better understanding of the heart dynamics and help it learn the essential dynamic representation within NODE.

2.4.3 Skip connections

Motivated by the strong performance of U-Net (Ronneberger et al., 2015) on medical images, where the skip connections between the contracting and expansive paths help retain high-resolution semantic features, we aim to preserve this property. However, direct skip connections from the encoder to the target decoder are not feasible due to the mismatch in time steps. As shown earlier, we achieve video interpolation during representation learning, so the expansive path for video generation can provide high-resolution semantic features when aligned with the target time steps. We introduce skip connections between the reconstruction decoder and the target decoder, allowing the learned video representation to enhance the performance of our target decoder.

2.4.4 Target decoder

In this work, our primary goal is the video segmentation of intracardiac structures, so the target predictions TS are the segmentation masks of classes containing background and other intracardiac structures (the number and types of intracardiac structures varies among different datasets, details are described in Section 2.6). Therefore, the target decoder consists of a sequence of upsampling convolutional layers similar to the reconstruction decoder and produce a multi-channel output corresponding to the probabilities of each class. An argmax function is applied to obtain the final segmentation masks. Since our proposed model is not limited to video segmentation, the target decoder can be adapted for different tasks.

2.5 Loss function

We apply the commonly used loss functions to verify the validity of our proposed framework. Other loss functions are also acceptable. The reconstruction loss is the mean square error:

R=MSEXsi,X̂si,

where X̂si is the reconstructed image by reconstruction decoder at si corresponding to the real image Xsi. The cross entropy loss and dice loss is applied to segmentation loss:

S=CEM,M̂+DiceM,M̂,

where M and M̂ are the ground truth and predicted segmentation mask respectively. The overall loss function is

=S+λR,

where λ is a hyper-parameter.

2.6 Datasets

In our experiments, we use a private dataset of PLAX videos and two public datasets, Echo-dynamic and CAMUS. The three datasets are described in detail below.

2.6.1 Private dataset

The resource of our private dataset comes from the Echonet-LVH (Duffy et al., 2022), which includes 12,000 labeled PLAX echocardiogram videos and expert annotations (measurements and calculations) to provide a baseline for studying the size of the cardiac chambers and the thickness of the wall. It lacks the segmentation masks of the cardiac structures. Because some samples of Echo-LVH are not of good quality, the videos are selected on the basis of video quality, which mainly takes into account the image size and clarity, the video recording method, and whether it is a PLAX video. Eventually, 2067 videos are selected. Afterward, each video is re-labeled on one or two frames, which are either ED frames or ES frames. The new annotations add labels for the segmentation of cardiac structures, including the contour of the LV, left atrium (LA), right ventricle (RV), left ventricular posterior wall (LVPW), interventricular septum (IVS) and right ventricular anterior wall (RVAW). In addition, more measurement labels are annotated, including LVPW thickness, LV internal dimension (LVID), IVS thickness, LA internal dimension, left ventricular outflow tract posterior dimension, right ventricular anterior wall thickness. Measurement labels annotate the coordinates of the starting and ending points and the length of the line. The labels do not necessarily exist on every annotated frame because some of those structures are not wholly visible in the image. Therefore, though the source video echocardiograms of this dataset is publicly available, the annotations being performed makes it private. All these works are conducted by an expert sonographer from Zhejiang Provincial People’s Hospital (Affiliated People’s Hospital), with more than 5 years of working experience.

2.6.2 EchoNet-dynamic

The EchoNet-Dynamic (Ouyang et al., 2019) includes 10,030 apical four-chamber view videos with corresponding labeled measurements, including ejection fraction, LV volume in ED and ES, and human expert tracings of the LV as an aid for studying machine learning approaches to evaluate cardiac function. It is the largest labeled medical video dataset made available publicly to researchers and provides a baseline for studying cardiac motion and chamber sizes.

2.6.3 CAMUS

The CAMUS dataset (Leclerc et al., 2019) is the largest publicly available fully annotated dataset for 2D echocardiographic assessment. The purpose of this dataset is to provide all materials to the community to solve the problem of echocardiographic image segmentation and volume estimation from 2D ultrasound sequences. It contains 2D apical four chamber (4CH) and two chamber (2CH) view sequences acquired from 500 patients, and the videos are annotated frame by frame. Segmentation masks include the LV, LA, and myocardium of the left ventricle (LVM).

2.7 Temporal consistency metrics

Usually, the intersection over union (IoU) and dice coefficient (Dice) are used to show the performance of segmentation and to measure the region similarity of predicted and label masks. However, temporal consistency of the video segmentation is also an important aspect in video segmentation since the evolution of object shapes is an important cue for relative downstream tasks. Stable temporal segmentation contributes to the automatic analysis of the heart, such as cardiac wall motion and many downstream tasks. Therefore, we utilize two metrics to measure temporal consistency.

From the perspective of the predicted video segmentation itself, we calculate the TC proposed by Perazzi et al. (2016). This method transforms a mask Mt in time t into polygons representing its contours P(Mt), then describes each point ptiP(Mt) using the Shape Context Descriptor (SCD) (Belongie et al., 2002). The matching of two masks is presented as Dynamic Time Warping (DTW) (Rabiner and Juang, 1993) problem that minimizes the SCD distances between the matched points while preserving the order in which the points are present in the shapes. The resulting mean cost per matched point is used as a measure of the temporal stability τ of this pair of segmentation. The smaller the metric, the better. In our paper, the mean mask matching of all adjacent frames is used to measure the temporal consistency of the whole video segmentation:

TC=1NTk=1Nt=1TτMkt1,Mkt,

where Mkt is the predicted mask of t-th frame in k-th video.

In terms of scenarios where the ground truth of the whole video is available, we propose a more intuitive metric based on the ground truth. Assuming that the labels are consistent and the predictions are also consistent and well aligned with the labels, then the dice of any adjacent segmentation of a consistent prediction should be very close. We define temporal consistency of dice as:

TCD=1NTk=1Nt=1T|DktDkt1|,

where Dkt is the dice similarity of t-th frame in k-th video. theoretically, TCD should be close to zero if the predictions are temporal consistent.

3 Results

3.1 Implementation details

Model design and training are done in Python using the PyTorch deep learning library on two NVIDIA Tesla P40 GPUs. The private dataset is split into train, valid and test subsets in the ratio of 8:1:1. These subsets of the other two public datasets are split by the publisher. Stochastic gradient descent (SGD) optimizer with a learning rate of 0.05 is used. We train Echo-ODE for 100 epochs with a batch size of 16. Random resize, random crop, random horizontal flip, color jitter, and random Gaussian blur are used for data augmentation. For hyperparameters, λ is set to 10 by empirical grid search over a focused range {1,5,10,20} that balance the magnitude scales of the two loss components. Dopri5 is chosen as our ODESolver. The time interval for integrating the NODE is set to [0,1], i.e., tn=1. Therefore, the timesteps of the input τ and output S is determined by the frame sampling rate. When frames are sampled at uniform intervals as it is the case in this paper, the timenodes are also uniformly spaced within the time interval [0, 1]. All of the experiments in this paper are trained from scratch. To show the effectiveness of our model, we compare the performance of Echo-ODE with the classical image-level model UNet Ronneberger et al. (2015) and the classical RNN-based video-level model ConvLSTM Hochreiter and Schmidhuber (1997) with the same loss functions. In order to compare the ability of modeling the dynamics of echocardiograms by LSTM cells and NODE, we replace the NODE blocks in Echo-ODE by LSTM cells to capture the temporal relationship of hidden states and abbreviate this method as Echo-LSTM. UNet is trained at the image level and applied to every frame during video prediction, just as most works have done so far. The inputs of ConvLSTM are the same as those of Echo-LSTM and Echo-ODE, which sample 4 images from the video every two frames around the annotated image. In the training stage, the labeled frame is randomly ranked in one of the 4 frames. In the validation and test stages, the labeled frame is set to be fixed in the last frame for repeatability of experiments. When we need to output the prediction of a full video, Echo-ODE gets all the 7 intermediate predictions of the clip by dense outputs, and ConvLSTM is carried out 2 times with each time a frame shift of the inputs to get all the predictions. The loss convergence behavior of the training process is shown in Supplementary Material. Training and validation loss of Echo-ODE demonstrate a smooth and steady decrease, reaching a stable plateau after approximately 80 epochs without signs of overfitting or underfitting. The reconstruction loss and segmentation loss have the same magnitude after scaled by the weight parameter λ, and the synchronized convergence of both losses suggests that the reconstruction and the segmentation mutually reinforce each other during joint training.

3.2 Performance on private dataset

In this section, we first verify that Echo-ODE is a valid method for representing heart dynamics. Then, we compare the segmentation performance of Echo-ODE with UNet, ConvLSTM, and Echo-LSTM in three aspects: region similarity, temporal consistency, and the ability to perform phase detection.

3.2.1 Dynamic representation

Figure 2 shows an example of video reconstruction and segmentation by Echo-ODE. It is a single run with video interpolation. As we can see, the textures of the original images are reconstructed perfectly except for the noise, and they are more likely to be the smoothed images. This is within our expectation because there are no skip connections from the encoder path and some details are missing. Since the images are reconstructed directly from the NODE, we can draw a tentative conclusion that the NODE model is largely capable of learning certain dynamic characteristics of the video and that NODE is a reliable method to model the dynamics of the heart.

Figure 2
Sequential ultrasound images of a heart, shown in three rows. The first and second row illustrate the original and reconstructed images of a video clip in grayscale. The third row displays the original images with color-coded segmentation highlighting various heart chambers and structures.

Figure 2. Examples of video reconstruction and segmentation by Echo-ODE. Top row: original images. Middle row: reconstructed images. Bottom row: segmentation masks.

3.2.2 Region similarity

The IoU and the dice are summarized in Table 1. As segmentation of LV is the most critical task in many applications, we list the dice of LV segmentation and the corresponding 95% confidence interval of these four methods in Table 2.

Table 1
www.frontiersin.org

Table 1. Comparison of segmentation performance of six cardiac structures. Higher values indicate better performance. The best results are highlighted in bold.

Table 2
www.frontiersin.org

Table 2. Dice and 95% confidence interval comparison of LV segmentation. The data format is Dice (confidence interval). Higher values indicate better performance. The best results are highlighted in bold.

Tables 1, 2 show that UNet, ConvLSTM, and Echo-ODE exhibited comparable performance in terms of the overall mean dice score. UNet demonstrates the best performance across most structures, but struggles with RVAW segmentation, which has fewer training instances. The dice score for Echo-ODE is slightly lower than that of UNet and ConvLSTM. We attribute the advantages of the latter two methods to two main factors. First, UNet is an image-level model that focuses on learning the ED and ES frames. The evaluation process is also carried out on these images, so UNet performs better in its specialized domain, while the other two methods allocate some of their focus to learning segmentation on images beyond just the ED and ES frames. Furthermore, both UNet and ConvLSTM utilize skip-connections from the encoder path, which combine low-level and high-level semantic features extracted by the model, providing richer information for the final output. In contrast, the Echo-ODE model directly decodes the segmentation results from the NODE output, leading to the loss of low-level information. However, our model achieves an accuracy comparable to that of the other two methods.

Compared to ConvLSTM, Echo-LSTM loses low-level semantic features from the encoder, resulting in lower segmentation performance (mean Dice: 0.718 vs. 0.726). Surprisingly, Echo-ODE achieves superior segmentation metrics despite similar feature limitations, outperforming ConvLSTM (mean Dice: 0.730 vs. 0.726) The structures of Echo-LSTM and Echo-ODE are the same except for the bottleneck, which suggests that NODE is more suitable than LSTM for modeling cardiac dynamics. This is because ODE learns the continuous derivation of echocardiograms, whereas LSTM just learns from discrete inputs.

Another potential strength of Echo-ODE is its ability to resist missed labels. As mentioned earlier, the segmentation is not fully labeled in our private dataset, and the RVAW is the least labeled, although the structure is partially visible in some of these images. Echo-ODE can deduce RVAW segmentation benefiting from a better dynamic representation and performs far better than UNet and ConvLSTM.

3.2.3 Temporal consistency

The results of TC are summarized in Table 3. Echo-ODE performs best in terms of temporal consistency in all six cardiac structures and is much more consistent than the other three methods. To provide a better visual perception, we use a continuous clip as an example and compare the segmentation results of the four methods in Figure 3. As shown, in the regions marked by dashed circles of the same color, the segmentation masks exhibit large fluctuations between adjacent frames, while Echo-ODE apparently performs better in these regions. Due to the lack of temporal context information, the predictions of UNet may be affected by image noise, resulting in large differences even when the changes between original images are visually subtle. Echo-ODE performs much better in terms of temporal consistency, which not only improves the TC by 20.28%, 21.61% and 18.68% compared to the other methods respectively, but also visually appears significantly more consistent.

Table 3
www.frontiersin.org

Table 3. Comparison of TC of video segmentation. Lower values indicate better performance. The best results are highlighted in bold.

Figure 3
A grid displays echocardiogram images overlaid with segmentation masks, comparing four methods: UNet, ConvLSTM, Echo-LSTM, and Echo-ODE. Each row represents one method, showing four sequential frames where dashed circles of different colors mark inconsistencies (large fluctuations) in segmentation results between adjacent frames.

Figure 3. Segmentation of a clip of neighbouring frames. Dashed circles of the same color mark inconsistencies (large fluctuations) of the segmentation results between adjacent frames.

3.2.4 Phase detection

To demonstrate that our temporally consistent model is more reliable in clinical scenarios, we test our model in downstream tasks. In a typical workflow for fully automated analysis of video echocardiograms, we first segment every frame of the video, then select the appropriate frames as ED and ES frames (phase detection) based on the segmentation results. Indicators such as the ejection fraction are then calculated to evaluate whether the heart is healthy. We conduct phase detection as an example to test whether the segmentation by Echo-ODE can effectively accomplish downstream tasks with better performance. The method follows the reference Zeng et al. (2023). All frame sequences Pn of the LV segmentation are extracted and a curve of the left ventricular area A is plotted as a function of time. The Savitzky-Golay filtering algorithm with filter width of 13 and polynomial order of three is applied to reduce noise interference. Finally, a peak detection function is used to identify peaks and valleys as ED and ES frames. This process is shown in Figure 4, and more details can be found in reference Zeng et al. (2023). Since our dataset is annotated with ED and ES frames by experts, these frames are regarded as the ground truth. A detected peak within 10 frames of the ground truth is considered a valid prediction. The mean absolute error of phase detection is listed in Table 4.

Figure 4
Illustration of the cardiac cycle phase detection process. The top of the figure displays a series of ultrasound images with color-coded segmentation masks. Below, a line graph illustrates left ventricular (LV) area changes over frames. The blue line represents Savitzky-Golay filtered LV area. Green vertical lines and red vertical lines represent the predicted ED frames and ES frames respectively. Red dots mark doctor-labeled frames. Legend indicates segmented LV area, filtered LV area, predicted frames, and labeled frames.

Figure 4. Illustration of the cardiac cycle phase detection process. The LV areas of a video along time was smoothed by Savitzky-Golay filtering algorithm, then the detected peaks and valleys of the curve were regarded as ED and ES frames.

Table 4
www.frontiersin.org

Table 4. Errors of cardiac cycle phase detection by different methods. Lower values indicate better performance. The best results are highlighted in bold.

As shown in Table 4, Echo-ODE has the least detection error in both ED frames, ES frames and overall compared to the other three methods. It shows that a consistent and stable segmentation of cardiac structures can be benifical in completing the phase detection task.

3.2.5 Robustness to arrhythmia

To demonstrate the robustness of our model in handling diverse cardiac conditions in clinical scenarios, we analyze the segmentation performance of patients with arrhythmia in the validation and test datasets, which include a total of 18 cases. The results are listed in Table 5. Compared to Table 1, Tables 3, 4, the performance of the four models has a decline in this scenario. The Dice score decreases by 0.015, 0.026, 0.030, and 0.011 for UNet, ConvLSTM, Echo-LSTM and Echo-ODE, respectively. Our proposed Echo-ODE achieves the least decrease and highest mean segmentation Dice score. It also performs the best in terms of temporal consistency and phase detection error.

Table 5
www.frontiersin.org

Table 5. Comparison of segmentation performance of patients with arrhythmia. : higher values indicate better performance. : lower values indicate better performance. The best results are highlighted in bold.

3.2.6 Other tasks

To test the efficiency of Echo-ODE in accomplishing other tasks, we replace the segmentation target decoder with a measurement decoder to measure the ventricular dimensions. The measurement decoder is similar to the original target decoder, but the output head predicts the probability heatmaps for the beginning and ending points of IVS thickness, LVID dimension, and LVPW thickness respectively, which is the TS in this scenario. The center of gravity of the output heatmaps is regarded as the points. At inference time, low-confidence pixels (with a score less than 0.3) are ignored when calculating the centroid. Ventricular dimensions are calculated using these points and the image resolution. Echo-ODE achieves an MAE of 1.0 mm (95% CI, 0.9–1.1 mm) for IVS thickness, 2.4 mm (95% CI, 2.1–2.7 mm) for LVID, and 1.4 mm (95% CI, 1.2–1.6 mm) for LVPW thickness, which shows very close performance to Duffy et al. (2022), where the MAEs are 1.2 mm, 2.4 mm, and 1.4 mm, respectively. These results demonstrate that Echo-ODE can perform the measurement task. An example in which we overlay the heatmaps on the echocardiography is displayed in Figure 5. The ground truth and predicted measurement lines are highlighted in red and cyan, respectively. As shown, the heatmaps are distributed along the boundaries and the annotated points also lie within the range of the heatmap with high confidence.

Figure 5
Ultrasound image showing cardiac structures with overlaid colored markers and lines indicating measurements. Red and cyan lines denote the measure lengths of ground truth and prediction, respectively. Blue dots mark the start and end points of the lines.

Figure 5. An example of measurement of ventricular dimensions. Red: label. Cyan: prediction. Blue: start/end point of the line.

3.3 Performance on public dataset

3.3.1 EchoNet-dynamic

The experiments are also carried out on the public EchoNet-Dynamic dataset with our model. We compare the results with the reference of Ouyang et al. (2020) who published the dataset, and another reference MAEF-Net by Zeng et al. (2023) who also conducted the tasks of segmentation and phase detection. Ouyang et al. perform an image-level semantic segmentation using the Deeplabv3 architecture. MAEF-Net uses a multi-attention mechanism to guide the network in capturing heartbeat features effectively and incorporates a deep supervision mechanism and spatial pyramid feature fusion to enhance feature extraction capabilities. The results are shown in Table 6.

Table 6
www.frontiersin.org

Table 6. Comparison of results on public dataset EchoNet-Dynamic. : higher values indicate better performance. : lower values indicate better performance. The best results are highlighted in bold.

We achieve better performance in dice similarity of segmentation than Ouyang et al., but slightly worse than Zeng. MAEF-Net uses a multi-attention mechanism to gather spatial and temporal information, which provides more low-level semantic features. As discussed earlier, Echo-ODE loses many low-level features, so the dice coefficient of segmentation struggles to surpass advanced 2D models. However, Echo-ODE performs better in the phase detection task. We can draw a preliminary conclusion that the prediction of Echo-ODE is more temporally stable and continuous, even though Zeng et al. do not report the temporal consistency of their segmentation.

3.3.2 CAMUS

Since CAMUS is fully annotated, we can evaluate performance not only on ED and ES frames, but also compare the temporal consistency of the predicted video segmentation with the ground truth. To this end, we conduct experiments on CAMUS. During training, only annotations of ED and ES frames are used, and testing is applied to the entire video. The results are summarized in Table 7, and the distributions of TCD for the three methods and the three structures are illustrated in Figure 6.

Table 7
www.frontiersin.org

Table 7. Comparison of performance on public dataset CAMUS. : higher values indicate better performance. : lower values indicate better performance. The best results are highlighted in bold.

Figure 6
A histogram grid of $TC_D$ compares U-Net, ConvLSTM, and Echo-ODE models across six categories: 2CH-LV, 2CH-LVM, 2CH-LA, 4CH-LV, 4CH-LVM, and 4CH-LA. Each row represents a different category, and each column represents a model. Bars show distribution patterns over specified ranges.

Figure 6. Distributions of TCD by (A) UNet, (B) ConvLSTM, (C) Echo-ODE.

As shown, the three methods demonstrate comparable overall accuracy, but Echo-ODE has much smaller TC and TCD than the other two methods. It is worth noting that the TC of the ground truth is 0.0812 and 0.0911 for the 2CH and 4CH views, respectively. The TC of Echo-ODE is very close to the ground truth, and the distributions of TCD by Echo-ODE are more concentrated near zero and have fewer significantly large errors. Clearly, Echo-ODE shows the best performance in temporal consistency.

3.4 Ablation study

As described in Section 2.4, we utilize a reconstruction block (Rec) on the baseline to enhance the dynamic representation of our model and add the skip connections (Skip) from the reconstruction path to the segmentation path. To show the effectiveness of these two blocks, we compare the results of these different methods and the results are listed in Table 8. Clearly, both Rec and Skip contribute to improvements in segmentation accuracy, but share similar temporal stability. This further confirms that the NODE framework has the property of producing temporally consistent predictions, regardless of how the decoders are designed.

Table 8
www.frontiersin.org

Table 8. Results of the ablation study. : higher values indicate better performance. : lower values indicate better performance. The best results are highlighted in bold.

3.5 Computational complexity

We compare the computational complexity among different models. The number of floating-point operations (#FLOPs) and the number of parameters (#Params) are obtained using the function thop.profile() in Python. In addition, we provide runtime analysis in our private test dataset. The results are summarized in Table 9. UNet is an image-level model, so the #FLOPs and #Params are smaller than those of the other three methods. Echo-LSTM has an extra reconstruction branch compared to ConvLSTM, so it has higher computational complexity. The only difference between Echo-ODE and Echo-LSTM is the bottleneck, so, as shown in the table, the #FLOPs of Echo-ODE are 1.19B greater than those of Echo-LSTM in our experimental setting. Since we use shared parameters in temporal modeling, the difference in #Params among the three video-level models is not particularly large. Due to the fewer parameters and single-frame input, UNet requires significantly less training time than other models. In Echo-ODE, the optimization of ODE blocks is more time consuming, resulting in a training time approximately twice that of LSTM-related models. The same conclusion can be drawn in terms of the inference time of single forward propagation (referring solely to the time of data flow through the model, excluding data reading and other precesses). It should be noted that a single forward propagation can produce 1, 4, 4, and 7 frames of segmentation for UNet, ConvLSTM, Echo-LSTM, and Echo-ODE respectively in our experimental setups. Thus, the time-consuming differences for video segmentation across these four models are not very large (1.22s, 1.26s, 1.56 and 1.95s respectively). The mean video length in the test dataset is 158. Therefore, although Echo-ODE has a larger model size and requires more training time, the inference time difference for long video segmentation is not significant (within 2 seconds). If the video is further sparsely sampled and input into similar trained Echo-ODE models, the prediction time for video segmentation will be further reduced. Considering the potential benefit of more temporally consistent video segmentation, it is worth the extra time.

Table 9
www.frontiersin.org

Table 9. Computational complexity of different methods.

4 Discussion

Segmentation of video echocardiograms plays an essential role in clinical diagnosis. Accurate and temporally consistent segmentation contributes to the reliable detection of ED and ES frames, quantification of cardiac functions, and other tasks. In this work, we propose a new framework, Echo-ODE, in which the heart is regarded as a dynamical system. We model the representation of dynamics using NODE. Echo-ODE learns the spatio-temporal relationships of the input video and outputs dense and continuous predictions. The main contributions of this paper are: (1) We propose Echo-ODE, an innovative framework that models the video echocardiogram as a dynamical system using NODE. The output of our model is dense and continuous. To our knowledge, this is the first work to apply NODE to echocardiography. (2) We achieve whole-heart video segmentation of the parasternal long-axis view (PLAX), including six cardiac structures, using our private dataset. (3) We introduce a new temporal consistency metric. (4) Experiments on both private and two public datasets demonstrate that Echo-ODE’s video segmentation is more temporally stable and consistent, which significantly contributes to improved performance in downstream tasks. Echo-ODE shows great potential to perform reliable and fully automatic video echocardiogram analysis.

The experimental results verify that Echo-ODE can learn the dynamic representation of the heart from video reconstruction and interpolation (Figure 2). Table 1 shows that Echo-ODE has a segmentation accuracy comparable to that of UNet and ConvLSTM. Since our model directly decodes the output of NODE without skip connections from the encoder path, it loses much low-level semantic information. Under this premise, it is encouraging to achieve comparable accuracy with the other methods. Echo-ODE performs better than Echo-LSTM, indicating that NODE is more suitable for modeling cardiac dynamic information than LSTM in our scenario. Additionally, a better segmentation accuracy for RVAW demonstrates that Echo-ODE has better video understanding and can more effectively deduce missing structures.

Table 3 shows that Echo-ODE video segmentation has the best temporal consistency in all cardiac structures. This advantage arises from the continuous dynamic representation of NODE, which contributes to the automatic analysis of downstream tasks. For example, the MAE of phase detection using LV segmentation by Echo-ODE is the smallest (Table 4).

Table 5 shows that although all models perform slightly worse in dealing with patients with arrhythmias, Echo-ODE is the least affected and achieves the best segmentation accuracy. UNet is the second least affected because it is an image-level model with minimal temporal correlation. Segmentation accuracy for LSTM-based methods experiences the largest decline due to their discrete modeling of video frames, which does not adapt well to temporal irregularity. Furthermore, when we change the decoder branch to the task of key-point detection for cardiac measurement, Echo-ODE performs well. Experiments on two public datasets, including views of 4CH and 2CH, also lead to similar conclusions (Tables 6, 7), demonstrating that our model can be applied to various sectional views. These experiments highlight the robustness of our proposed model in handling various tasks and cardiac conditions in clinical scenarios. The ablation study (Table 8) shows that both the reconstruction block and the skip connection block in our model contribute to better performance.

Table 9 shows the computational complexity of the four models. Echo-ODE involves temporally continuous calculations, so its #FLOPs are higher than the other three methods. However, the #Params are not significantly larger because all these video-level methods use shared parameters in temporal modeling. Although Echo-ODE requires more time for a single forward propagation, it can predict nearly twice as many frames of segmentation. Considering the dense output scheme of the differential equation solver, Echo-ODE enables more continuous predictions without additional computational resources or time, making it more computationally efficient, so it is worth the extra time on the premise of its real-time prediction.

Our proposed Echo-ODE has several natural advantages. Firstly, it is based on the biodynamical modeling of specific scenarios, as explained in the introduction chapter. This not only strengthens the modeling ability, but also enhances the interpretability of the network. It offers a more functional alternative to optical flow. Secondly, the input of Echo-ODE does not need to have a constant time duration. It can better adapt to data collected from multicenter and multimachine setups, making it more suitable for transfer learning or multicenter learning. Last but not least, the continuity of the output can be utilized for super-resolution in the time dimension. For example, applying Echo-ODE to M-mode echocardiography can produce the corresponding vector graph and contribute to more accurate diagnoses.

However, this work has limitations. Our model and experiments leave much room for further exploration. (1) We need to gain a deeper insight into the hidden space and the derivative function to explore the interpretability of the model, and how are they connected to the biomechanics or underlying dynamical system. (2) Finding a way to reduce the loss of low-level semantic information is crucial. Although our reconstruction block strengthens the learning of dynamic representation, it cannot yet compensate for the loss of detailed information. (3) The computational cost of Echo-ODE is relatively higher than that of other mainstream CNN models. (4) We seek a more explainable method to integrate overall dynamic information into the neural derivative function. (5) The core experimental comparison in this work is deliberately focused on Echo-ODE (continuous modeling) against the most direct and conceptually relevant RNN models (discrete modeling). Providing a broader performance comparison across different methodological families is a planned focus of our immediate future work (6) The generalizability of Echo-ODE on more datasets, other echocardiographic views, patient cohorts and clinical relevance should be further validated. (7) We can explore the impact of various ODE solvers on the performance of this model. (8) The parameter settings of Savitzky-Golay filter used in the phase detection task is rigid. Considering that it may be dependent on factors such as recording frame rates or fibrillation, future work will focus on dynamically adapting filter parameters for a more generalizable implementation.

5 Conclusion

In this work, we propose Echo-ODE, an NODE-based network that learns the biodynamics of video echocardiograms. It achieves promising performance on both private and public datasets. The results demonstrate that the temporal stability and consistency of the predictions by Echo-ODE have great potential to enable reliable, fully automatic echocardiogram analysis.

Data availability statement

The public data used in this study is available in the article. The private data are not publicly available because of privacy and security concerns, but are available upon reasonable request from the corresponding author.

Author contributions

WL: Formal Analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review and editing. YW: Investigation, Software, Writing – original draft, Writing – review and editing. WD: Investigation, Writing – original draft. YW: Data curation, Investigation, Writing – original draft. HX: Conceptualization, Investigation, Supervision, Writing – original draft, Writing – review and editing. DK: Conceptualization, Project administration, Writing – original draft.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work is supported in part by the National Natural Science Foundation of China (Grant nos. 12090020 and 12090025), the R&D project of Pazhou Lab (Huangpu) (Grant nos. 2023K0605) and the Medical Science and Technology Project of Zhejiang Province (Grant nos. 2025KY564).

Conflict of interest

The authors declare that the research is conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fphys.2025.1629121/full#supplementary-material

References

Akbari S., Tabassian M., Pedrosa J., Queiros S., Papangelopoulou K., D’Hooge J. (2024). Beas-net: a shape-prior-based deep convolutional neural network for robust left ventricular segmentation in 2-d echocardiography. IEEE Trans. Ultrasonics, Ferroelectr. Freq. Control 71, 1565–1576. doi:10.1109/TUFFC.2024.3418030

PubMed Abstract | CrossRef Full Text | Google Scholar

Alsharqi M., Woodward W., Mumith J., Markham D., Upton R., Leeson P. (2018). Artificial intelligence and echocardiography. Echo Res. and Pract. 5, R115–R125. doi:10.1530/ERP-18-0056

PubMed Abstract | CrossRef Full Text | Google Scholar

Ballas N., Yao L., Pal C., Courville A. (2015). Delving deeper into convolutional networks for learning video representations. arXiv preprint arXiv:1511.06432.

Google Scholar

Belongie S., Malik J., Puzicha J. (2002). Shape matching and object recognition using shape contexts. IEEE Trans. Pattern Analysis Mach. Intell. 24, 509–522. doi:10.1109/34.993558

CrossRef Full Text | Google Scholar

Chen R. T., Rubanova Y., Bettencourt J., Duvenaud D. K. (2018). Neural ordinary differential equations. Adv. Neural Inf. Process. Syst. 31. doi:10.48550/arXiv.1806.07366

CrossRef Full Text | Google Scholar

Duffy G., Cheng P. P., Yuan N., He B., Kwan A. C., Shun-Shin M. J., et al. (2022). High-throughput precision phenotyping of left ventricular hypertrophy with cardiovascular deep learning. JAMA Cardiol. 7, 386–395. doi:10.1001/jamacardio.2021.6059

PubMed Abstract | CrossRef Full Text | Google Scholar

Farsalinos K. E., Daraban A. M., Ünlü S., Thomas J. D., Badano L. P., Voigt J.-U. (2015). Head-to-head comparison of global longitudinal strain measurements among nine different vendors: the eacvi/ase inter-vendor comparison study. J. Am. Soc. Echocardiogr. 28, 1171–e2. doi:10.1016/j.echo.2015.06.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Guo L., Lei B., Chen W., Du J., Frangi A. F., Qin J., et al. (2021). Dual attention enhancement feature fusion network for segmentation and quantitative analysis of paediatric echocardiography. Med. Image Anal. 71, 102042. doi:10.1016/j.media.2021.102042

PubMed Abstract | CrossRef Full Text | Google Scholar

Hasan M. K., Yang G., Yap C. H. (2025). Motion-enhancement to echocardiography segmentation via inserting a temporal attention module: an efficient, adaptable, and scalable approach. arXiv preprint arXiv:2501.14929.

Google Scholar

Hochreiter S., Schmidhuber J. (1997). Long short-term memory. Neural Comput. 9, 1735–1780. doi:10.1162/neco.1997.9.8.1735

PubMed Abstract | CrossRef Full Text | Google Scholar

Hu J., Xue W., Cheng J., Liu Y., Zhuo W., Ni D. (2025). “Echoone: segmenting multiple echocardiography planes in one model,” in Proceedings of the computer vision and pattern recognition conference, 5207–5216.

Google Scholar

Huang H., Nijjar P. S., Misialek J. R., Blaes A., Derrico N. P., Kazmirczak F., et al. (2017). Accuracy of left ventricular ejection fraction by contemporary multiple gated acquisition scanning in patients with cancer: comparison with cardiovascular magnetic resonance. J. Cardiovasc. Magnetic Reson. 19, 34–39. doi:10.1186/s12968-017-0348-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Koh A. S., Tay W. T., Teng T. H. K., Vedin O., Benson L., Dahlstrom U., et al. (2017). A comprehensive population-based characterization of heart failure with mid-range ejection fraction. Eur. J. Heart Fail. 19, 1624–1634. doi:10.1002/ejhf.945

PubMed Abstract | CrossRef Full Text | Google Scholar

Lane E. S., Azarmehr N., Jevsikov J., Howard J. P., Shun-Shin M. J., Cole G. D., et al. (2021). Multibeat echocardiographic phase detection using deep neural networks. Comput. Biol. Med. 133, 104373. doi:10.1016/j.compbiomed.2021.104373

PubMed Abstract | CrossRef Full Text | Google Scholar

Lang R. M., Badano L. P., Mor-Avi V., Afilalo J., Armstrong A., Ernande L., et al. (2015). Recommendations for cardiac chamber quantification by echocardiography in adults: an update from the american society of echocardiography and the european association of cardiovascular imaging. Eur. Heart Journal-Cardiovascular Imaging 16, 233–270. doi:10.1093/ehjci/jev014

PubMed Abstract | CrossRef Full Text | Google Scholar

Leclerc S., Smistad E., Pedrosa J., Østvik A., Cervenansky F., Espinosa F., et al. (2019). Deep learning for segmentation using an open large-scale dataset in 2d echocardiography. IEEE Trans. Med. Imaging 38, 2198–2210. doi:10.1109/TMI.2019.2900516

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu F., Wang K., Liu D., Yang X., Tian J. (2021). Deep pyramid local attention neural network for cardiac structure segmentation in two-dimensional echocardiography. Med. Image Anal. 67, 101873. doi:10.1016/j.media.2020.101873

PubMed Abstract | CrossRef Full Text | Google Scholar

Loehr L. R., Rosamond W. D., Chang P. P., Folsom A. R., Chambless L. E. (2008). Heart failure incidence and survival (from the atherosclerosis risk in communities study). Am. J. Cardiol. 101, 1016–1022. doi:10.1016/j.amjcard.2007.11.061

PubMed Abstract | CrossRef Full Text | Google Scholar

Ma Q., Li L., Robinson E. C., Kainz B., Rueckert D., Alansary A. (2022). Cortexode: learning cortical surface reconstruction by neural odes. IEEE Trans. Med. Imaging 42, 430–443. doi:10.1109/TMI.2022.3206221

PubMed Abstract | CrossRef Full Text | Google Scholar

Malm S., Frigstad S., Sagberg E., Larsson H., Skjaerpe T. (2004). Accurate and reproducible measurement of left ventricular volume and ejection fraction by contrast echocardiography: a comparison with magnetic resonance imaging. J. Am. Coll. Cardiol. 44, 1030–1035. doi:10.1016/j.jacc.2004.05.068

PubMed Abstract | CrossRef Full Text | Google Scholar

Moradi S., Oghli M. G., Alizadehasl A., Shiri I., Oveisi N., Oveisi M., et al. (2019). MFP-Unet: a novel deep learning based approach for left ventricle segmentation in echocardiography. Phys. Medica 67, 58–69. doi:10.1016/j.ejmp.2019.10.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Ouyang D., He B., Ghorbani A., Lungren M. P., Ashley E. A., Liang D. H., et al. (2019). “Echonet-dynamic: a large new cardiac motion video data resource for medical machine learning,” in NeurIPS ML4H Workshop, Vancouver, BC, Canada. 5.

Google Scholar

Ouyang D., He B., Ghorbani A., Yuan N., Ebinger J., Langlotz C. P., et al. (2020). Video-based AI for beat-to-beat assessment of cardiac function. Nature 580, 252–256. doi:10.1038/s41586-020-2145-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Özçelik Y. B., Altan A. (2023). Overcoming nonlinear dynamics in diabetic retinopathy classification: a robust ai-based model with chaotic swarm intelligence optimization and recurrent long short-term memory. Fractal Fract. 7, 598. doi:10.3390/fractalfract7080598

CrossRef Full Text | Google Scholar

Park S., Kim K., Lee J., Choo J., Lee J., Kim S., et al. (2021). Vid-ODE: continuous-time video generation with neural ordinary differential equation. Proc. AAAI Conf. Artif. Intell. 35, 2412–2422. doi:10.1609/aaai.v35i3.16342

CrossRef Full Text | Google Scholar

Perazzi F., Pont-Tuset J., McWilliams B., Van Gool L., Gross M., Sorkine-Hornung A. (2016). “A benchmark dataset and evaluation methodology for video object segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27-30 June 2016, 724–732. doi:10.1109/cvpr.2016.85

CrossRef Full Text | Google Scholar

Pontryagin L. S. (1985). The mathematical theory of optimal processes and differential games. Tr. Mat. Instituta Im. VA Steklova 169, 119–158. doi:10.2307/2312867

CrossRef Full Text | Google Scholar

Rabiner L., Juang B.-H. (1993). Fundamentals of speech recognition. Prentice-Hall, Inc.

Google Scholar

Ronneberger O., Fischer P., Brox T. (2015). “U-net: convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-assisted Intervention, Munich, Germany, October 5-9, 2015, 234–241. doi:10.1007/978-3-319-24574-4_28

CrossRef Full Text | Google Scholar

Rubanova Y., Chen R. T., Duvenaud D. K. (2019). Latent ordinary differential equations for irregularly-sampled time series. Adv. Neural Inf. Process. Syst. 32. doi:10.48550/arXiv.1907.03907

CrossRef Full Text | Google Scholar

Shampine L. F., Jay L. O. (2015). “Dense output,” in Encyclopedia of applied and computational mathematics (Springer Berlin Heidelberg), 339–345.

CrossRef Full Text | Google Scholar

Su C., Zhou Y., Ma J., Chi H., Jing X., Jiao J., et al. (2024). Janet: a joint attention network for balancing accuracy and speed in left ventricular ultrasound video segmentation. Comput. Biol. Med. 169, 107856. doi:10.1016/j.compbiomed.2023.107856

PubMed Abstract | CrossRef Full Text | Google Scholar

Ta K., Ahn S. S., Thorn S. L., Stendahl J. C., Zhang X., Langdon J., et al. (2024). Multi-task learning for motion analysis and segmentation in 3d echocardiography. IEEE Trans. Med. Imaging 43, 2010–2020. doi:10.1109/TMI.2024.3355383

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu H., Liu J., Xiao F., Wen Z., Cheng L., Qin J. (2022). Semi-supervised segmentation of echocardiography videos via noise-resilient spatiotemporal semantic calibration and fusion. Med. Image Anal. 78, 102397. doi:10.1016/j.media.2022.102397

PubMed Abstract | CrossRef Full Text | Google Scholar

Yildiz C., Heinonen M., Lahdesmaki H. (2019). ODE2VAE: deep generative second order ODEs with bayesian neural networks. Adv. Neural Inf. Process. Syst. 32. doi:10.48550/arXiv.1905.10994

CrossRef Full Text | Google Scholar

Zeng Y., Tsui P.-H., Pang K., Bin G., Li J., Lv K., et al. (2023). MAEF-Net: Multi-attention efficient feature fusion network for left ventricular segmentation and quantitative analysis in two-dimensional echocardiography. Ultrasonics 127, 106855. doi:10.1016/j.ultras.2022.106855

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang J., Gajjala S., Agrawal P., Tison G. H., Hallock L. A., Beussink-Nelson L., et al. (2018). Fully automated echocardiogram interpretation in clinical practice: feasibility and diagnostic accuracy. Circulation 138, 1623–1635. doi:10.1161/CIRCULATIONAHA.118.034338

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: echocardiogram, cardiac segmentation, temporal consistency, physical dynamics representation, neural ordinary differential equations

Citation: Lu W, Wang  Y, Dai  W, Wu  Y, Xu H and Kong D (2025) Echo-ODE: A dynamics modeling network with neural ODE for temporally consistent segmentation of video echocardiograms. Front. Physiol. 16:1629121. doi: 10.3389/fphys.2025.1629121

Received: 02 June 2025; Accepted: 30 July 2025;
Published: 18 August 2025.

Edited by:

Wufeng Xue, Shenzhen University, China

Reviewed by:

Joakim Sundnes, Simula Research Laboratory, Norway
Junjie Hu, Sichuan University, China
Brett A Meyers, Purdue University, United States

Copyright © 2025 Lu, Wang , Dai , Wu , Xu and Kong. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Dexing Kong, ZHhrb25nQHpqdS5lZHUuY24=; Hao Xu, aGFvLnh1QHpqbnUuZWR1LmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.