Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Med., 29 August 2025

Sec. Precision Medicine

Volume 12 - 2025 | https://doi.org/10.3389/fmed.2025.1659122

This article is part of the Research TopicArtificial Intelligence in Imaging, Pathology, and Genetic Analysis of Brain Tumor in the Era of Precision MedicineView all 4 articles

Single-scan adaptive graph filtering for dynamic PET denoising by exploring intrinsic spatio-temporal structure


Shiyao GuoShiyao Guo1Xiaopeng LiXiaopeng Li1Shengting PanShengting Pan2Dan Zhang
Dan Zhang1*
  • 1School of Big Data and Statistics, Guizhou University of Finance and Economics, Guiyang, China
  • 2Audit Office, Guizhou University of Finance and Economics, Guiyang, China

The performance of Dynamic Positron Emission Tomography (PET) is often degraded by high noise levels. A key challenge is the significant variability across scans, which makes fixed denoising models suboptimal. Furthermore, current denoising algorithms are often confined to a single data domain, limiting their ability to capture deeper structural information. To overcome these limitations, we introduce a novel single-scan adaptive spatio-temporal graph filtering (ST-GF) technique. The fundamental principle is to explore the latent structure of the data by representing it in a graph-signal space. Unlike deep learning approaches requiring external training data, our algorithm works directly on a single acquisition. It maps the noisy sinogram to a graph to reveal its underlying spatio-temporal structure—such as spatial similarities and temporal correlation—that is obscured by noise in the original domain. The core of the framework is an iterative process where a graph filter is adaptively constructed based on this latent structure. This ensures the denoising operation is precisely tailored to the unique characteristics of the single scan being processed, effectively separating the true signal from noise. Experiments on simulated and in vivo datasets show our approach delivers superior performance. By leveraging the latent structure found exclusively within each scan, our method operates without prior training and remains immune to potential biases or interference from irrelevant external data. This self-contained approach grants ST-GF high robustness and flexibility, highlighting its substantial potential for practical applications.

1 Introduction

Dynamic Positron Emission Tomography (PET) provides a valuable, noninvasive tool for diagnosing a range of neurodegenerative disorders, such as Alzheimer's disease (13). By monitoring the temporal distribution of tracers in living tissues over time in vivo, PET provides valuable quantitative information about biological and physiological processes (4). However, a primary challenge in dynamic PET is that the short acquisition time required for high temporal resolution leads to low photon counts per frame (5). This results in reconstructed images of noise and low-quality, which impact the reliability of clinical diagnosis and quantitative analysis.

Pre-denoising data in the sinogram domain represents one of the effective methods for improving reconstructed image quality. This is because the noise in sinograms, which follows a predictable Poisson distribution (6, 7), is statistically easier to handle than the complex, spatially-correlated noise in the final image domain. Consequently, much of the foundational work in this area has focused on adapting spatial filtering techniques from natural image processing. Prominent examples include methods based on Non-Local Means (NLM) (8) and Block-Matching 3D filtering (BM3D) (9), as well as sinogram-based dynamic image guided filtering (SDIGF) (10). A primary limitation of these methods, however, is that they process each dynamic frame independently, thereby ignoring the critical temporal correlations between frames, which poses a major limitation, particularly in scenarios involving low counts. More recently, data-driven deep learning methods, particularly Transformer-based architectures, have shown strong performance by learning complex features from data (11, 12). Despite their effectiveness, they introduce new challenges. Their reliance on large, external training datasets is a primary concern, not only due to data scarcity but because the generic patterns learned from a database may not match the unique characteristics of a specific scan, leading to potential biases and suboptimal results. Furthermore, their “black-box” nature poses issues for clinical interpretability and trust, as studies have shown that while AI can preserve relative texture information better than standard filters, it can also significantly alter the quantitative values of these features, impacting downstream analysis and diagnosis (13).

Beyond the specific limitations of each methodology, a more fundamental challenge persists: many existing frameworks are optimized for the intermediate sinogram data, not the final reconstructed image. Given that image reconstruction is a complex inverse problem, an optimally denoised sinogram does not guarantee an optimally reconstructed image if critical structural information is lost in the denoising process. Therefore, an ideal framework for denoising dynamic sinogram must not only leverage complex spatio-temporal correlations but also be directly optimized for the final image quality.

In the context, Graph Signal Processing (GSP) (14, 15) emerges as a particularly promising paradigm for addressing such challenges. GSP has gained prominence as an effective method for analyzing signals on complex, irregular structures, with applications in reconstruction (16, 17) and denoising (18). Unlike classical filters that use a fixed basis (like Fourier or DCT), graph filtering is data-adaptive; its basis functions are derived directly from the relationships within the data itself. This structure-adaptive property allows graph filters to better capture the underlying signal's true characteristics. Building on these principles, a previous study by the authors developed a kernel-based graph filter (19), which demonstrated the potential of this approach but was itself limited by high computational complexity and an exclusive focus on temporal correlations, neglecting spatial information.

To address these shortcomings, this paper introduces an advanced single-scan adaptive spatio-temporal graph filtering (ST-GF) framework. The core principle of our framework is to progressively uncover the latent spatio-temporal structure inherent within the scan itself. This is achieved through a unique iterative refinement process where, instead of applying a static filter, the framework reconstructs the graph filter in each step from the most recent, denoised version of the signal. This strategy ensures that the filter dynamically adapts to the underlying signal structure as it becomes clearer, dramatically improving its ability to separate signal from noise. Furthermore, the novel dual-domain stopping criterion directly addresses the fundamental issue of optimizing for the incorrect target. By incorporating feedback from the reconstructed image, this criterion ensures the optimization process is explicitly oriented toward enhancing final image quality, not merely intermediate data metrics. The key contributions of this paper are the design of this new framework and its extensive evaluation, demonstrating its effectiveness, particularly in challenging low-count cases.

The paper is organized as follows: Section 2 introduces the proposed methodology. Sections 3 and 4 present the experimental results. Finally, Sections 5 and 6 provide a discussion and a concluding summary.

2 Materials and methods

2.1 Preliminaries

2.1.1 Dynamic PET imaging

In PET, the scanner detects pairs of photons along Lines of Response (LORs) that pass through the body. These detection events are systematically organized into a two-dimensional data plot known as a sinogram (20). In dynamic PET, a sequence of such sinograms, {pj}j=1F, is acquired over time to capture the tracer's kinetic behavior, where j∈{1, …, F} serves as the index for each of the F total frames.

The measurements for each dynamic frame j are statistically characterized by the Poisson distribution. The measured sinogram pjms is typically treated as a collection of independent Poisson random variables, with the likelihood function expressed by Pain et al. (21) and Qi and Leah (22):

P(pj|xj)=i=1mse-(p̄j)i(p̄j)i(pj)i(pj)i!,    (1)

where p̄j represents the expected (mean) sinogram for an underlying tracer distribution image xjn. The low counts inherent in the short acquisition times required for high temporal resolution mean that the measured data pj is inevitably corrupted by significant noise governed by this statistical model.

The ultimate goal of PET imaging is to obtain a high-quality image sequence {xj}j=1F. These images are computationally reconstructed from the noisy sinogram data, often by maximizing the Poisson log-likelihood. A widely used algorithm for this task is the Maximum-Likelihood Expectation Maximization (MLEM) method (22):

xjt+1=xjt(HT(pj(Hxjt+rj)))(HT1n).    (2)

As shown in Equation 2, the quality of the resulting reconstructed image xj is directly reliant on quality of input sinogram pj. The Poisson noise, as modeled in Equation 1, propagates and is often amplified during the reconstruction process, leading to final images with poor signal-to-noise ratios.

Therefore, an effective denoising of the sinograms is crucial for high-quality dynamic PET imaging. A conventional strategy is to apply a spatial denoising filter to each sinogram frame pj independently before reconstruction. However, this frame-by-frame denoising approach has a fundamental limitation: the dynamic sinogram frames are not independent. They exhibit strong temporal correlation due to the gradual kinetics of the tracer, and substantial spatial similarity as they represent the same underlying anatomy.

These inherent spatio-temporal dependencies within the sinogram data provide powerful prior information. By treating each sinogram as an isolated dataset, conventional denoising methods fail to leverage this vital inter-frame information, leading to suboptimal performance, especially in challenging low-count cases. To address this critical gap, our work introduces a novel graph-based filtering framework designed to explicitly uncover and model the latent structure defined by these intrinsic correlations, enabling superior noise suppression.

2.1.2 Graph-based signal filtering

A weighted undirected graph G(V, ε, W) is a flexible and powerful representation for modeling signal relationships in image processing. In this context, each vertex viV = {v1, …, vn} corresponds to a pixel or voxel. The edges ε define the connectivity between these vertices, and W ∈ ℝn×n is the weighted adjacency matrix, where wij quantifies the similarity between vi and vj (23). When applied to a signal on the graph, the adjacency matrix performs a weighted averaging of neighboring node values, giving it an inherent low-pass characteristic (24).

The Graph Laplacian L is defined as L = DW, where D = diag(d1, …, dn) is the degree matrix with di=jwij. In contrast to the smoothing nature of W, the Laplacian emphasizes the differences between connected nodes [i.e., (Lx)i=jwij(xi-xj)], thereby exhibiting a natural high-pass characteristic (24). The spectral decomposition of the Laplacian,

L=UΛUT,    (3)

provides the basis for graph spectral analysis. Here, U contains the eigenvectors and Λ = diag(λ1, …, λn) contains the corresponding eigenvalues, sorted such that 0 = λ1 ≤ ⋯ ≤ λn.

This spectral decomposition underpins graph filtering. The eigenvalues of L are interpreted as graph frequencies. Eigenvectors associated with small eigenvalues (low frequencies) vary slowly across the graph and represent the signal's smooth, principal components. Conversely, eigenvectors for large eigenvalues (high frequencies) oscillate rapidly and typically correspond to noise or fine details (14). Therefore, an effective strategy for noise suppression is to apply a low-pass filter that attenuates the high-frequency components while preserving the essential low-frequency structures.

Building on these principles, our methodology is designed to suppress noise by transforming the data into a structure-aware domain where these relationships are made explicit. We select the adjacency matrix W to serve as a graph low-pass filter, as its inherent smoothing properties respect the underlying data structure it represents. To clarify the principle of graph-based low-pass filtering that underpins our method, its mechanism can be understood from the following two perspectives.

From an intuitive weighted averaging perspective, the essence of graph filtering is to replace each pixel's value with a weighted combination of a set of pixels with highly similar feature vectors. Unlike a traditional Gaussian filter, which relies on fixed spatial proximity, the weights in a graph filter shown in Equation 7 are determined by data-driven similarity. This mechanism effectively smooths out random, isolated noise points during the averaging process because their feature vectors differ significantly from those of their neighbors in the true structure. Meanwhile, the feature-similar pixels that constitute the image's structure are collectively preserved due to the high weights between them.

From the more rigorous perspective of Graph Spectral Filtering theory, this process has a solid mathematical foundation. The eigenvalues of the Graph Laplacian L are interpreted as graph frequencies, where small eigenvalues (low frequencies) correspond to the smooth, structural components of the image, and large eigenvalues (high frequencies) correspond to noise. Our use of the adjacency matrix W as a filter is mathematically equivalent to attenuating the signal components associated with high frequencies in the graph spectral domain. Thus, graph filtering achieves effective suppression of noise (high-frequency) and precise preservation of structure (low-frequency) by exploiting the intrinsic similarities within the data.

This filtering operation is performed in both spatial and temporal dimensions by constructing separate graphs to model anatomical consistency within frames and tracer kinetics across frames. The core of our proposed framework, detailed in the following sections, is an iterative process that refines not only the signal but also the estimation of its latent structure simultaneously.

2.2 Methods

The core philosophy of our proposed method is to perform denoising by uncovering the latent spatio-temporal structure inherent within a single, noisy dynamic PET scan. Unlike learning-based methods that rely on large, external datasets, our approach is single-scan adaptive. It operates self-sufficiently on the data at hand, thus remaining immune to the biases and variabilities of external datasets. To achieve this, this paper proposes an iterative filtering framework where the estimated signal and its underlying graph structure are progressively and jointly refined, as summarized in Algorithm 1 and illustrated in the flowchart in Figure 1.

Algorithm 1
www.frontiersin.org

Algorithm 1. The proposed iterative spatio-temporal graph filtering (ST-GF) algorithm.

Figure 1
Flowchart illustrating a denoising process for dynamic PET sinograms. It shows steps involving temporal correlation, spatial similarity, and dual-domain stopping criterion. Outputs include denoised sinograms through temporal and spatial graph filters.

Figure 1. Flowchart of the proposed iterative spatio-temporal graph filtering (ST-GF) framework. At the start of each iteration (i), the current sinogram data (Y(i)) is used to construct an adaptive spatial filter (FS(i)) and temporal filter (FT(i)). These filters are then applied to produce the refined sinogram Y(i+1). This output then serves as the input for constructing the filters in the next iteration, creating a feedback loop that progressively suppresses noise until a dual-domain stopping criterion is met.

2.2.1 Iterative filtering framework and filter construction

Let the original noisy sinogram data be denoted as Y(0), and the denoised result after the i-th iteration as Y(i). The update rule for each iteration is given by:

Y(i+1)=FS(i)Y(i)(FT(i))T    (4)

where FS(i) and FT(i) are the spatial and temporal graph filters at iteration i.

This iterative update is the heart of our method. Here, the graph filters are not static; they represent the algorithm's current estimate of the latent data structure and are dynamically reconstructed based on the cleaner signal Y(i). This process creates a powerful bootstrapping feedback: using an imperfect structure to find a better signal, which in turn is used to define a more precise structure. It is this mechanism that allows the framework to converge on the true underlying signal structure with increasing fidelity.

Specifically, the filters are defined as symmetrically normalized adjacency matrices:

FS(i)=(Ds(i))-1/2Ws(i)(Ds(i))-1/2,FT(i)          =(Dt(i))-1/2Wt(i)(Dt(i))-1/2    (5)

where the adjacency matrices Ws(i) and Wt(i) encode the spatial and temporal similarities, respectively. These matrices are re-computed at each iteration i based on the current denoised data Y(i). The corresponding degree matrices are Ds(i) and Dt(i). The specific methods for constructing Ws(i) and Wt(i) are detailed in the following sections.

2.2.2 Spatial graph filter design

To capture and define the latent anatomical structure at each iteration (k), the spatial adjacency matrix Ws(k) is constructed as follows. To ensure this matrix is robust against noise while adapting to the progressively cleaner signal, we first generate composite frames Zj(k) by fusing adjacent frames from the current denoised image sequence Y(k−1):

Zj(k)=lN(j)αlYl(k-1)    (6)

where N(j) denotes the set of frame indices involved in generating the composite frame Zj(k), and αl are normalized weights.

After generating these iteration-specific composite frames, the spatial adjacency matrix Ws(k) is constructed by computing the similarity between any two image patches (nodes) m and n:

(Ws(k))mn=exp(-||Zm(k)-Zn(k)||2σs2)    (7)

where σs is a parameter that controls the sensitivity to spatial differences.

2.2.3 Temporal graph filter design

Similarly, to model the latent structure of the tracer kinetics at iteration (k), the temporal adjacency matrix Wt(k) is recomputed at each step to capture the evolving characteristics of the time-activity curves (TACs). This matrix is computed based on the similarity between entire frames in the current denoised image sequence Y(k−1):

(Wt(k))pq=exp(-||Yp(k-1)-Yq(k-1)||2σt2)    (8)

where σt is a parameter controlling sensitivity to temporal changes. This iterative construction of the temporal filter ensures that it adaptively enforces smoothness along the time-activity curves (TACs), more effectively suppressing temporal noise while preserving the true metabolic trends revealed in the progressively cleaner signal.

2.2.4 Stopping criterion

A critical component of any iterative framework determines when to terminate the process. A naive criterion based only on the convergence of the sinogram data (Y) might not yield the best possible reconstructed image. To address this, we introduce a dual-domain stopping criterion that is guided by the final image quality. The iterative process stops when the weighted relative change in both the sinogram domain (Y) and the image domain (I) falls below a predefined tolerance ϵ:

W1Y(i+1)-Y(i)FY(i)F+W2I(i+1)-I(i)FI(i)F<ϵ    (9)

where I(i+1) is the image reconstructed from the updated sinogram Y(i+1), ∥·∥F denotes the Frobenius norm, and the weights w1 and w2 balance the two terms. By incorporating feedback from the reconstructed image, this criterion provides a more reliable path toward achieving a high-quality and well-balanced result.

3 Results

3.1 Simulation setup

Dynamic PET scans were simulated to mimic data acquisition from a GE DST whole-body scanner (25). The simulation was based on a 111 × 111 pixel Zubal head phantom incorporating background, gray matter, white matter, and a central 5 × 5 pixel lesion (Figure 2a). The temporal dynamics were established using regional time-activity curves produced by a three-compartment model (26), an example of which is shown in Figure 2b. Parameters for this kinetic model were adopted from the study (27). These activities were then sampled over a 24-frame dynamic scan sequence with progressively longer durations (4 × 20 s, 4 × 40 s, 4 × 60 s, 4 × 180 s, and 8 × 300 s).

Figure 2
(a) A grayscale brain scan highlighting a bright spot representing a tumor on the right side. (b) A graph showing activity levels over time for gray matter, white matter, and a tumor, with the tumor showing the highest activity. (c) Another grayscale brain scan similar to (a), with the tumor indicated on the left side.

Figure 2. (a) The phantom used for simulation, (b) The time-activity curve, (c) The PET activity image.

Our simulation process began by constructing the ground-truth activity distribution: we populated the anatomical regions of the phantom with their respective time-activity curves (TACs). We then executed a forward projection by employing the system matrix available in the Fessler toolbox to compute noise-free sinograms. Subsequently, noise was incorporated in two stages: first, a 20% background of random and scatter events was added; second, Poisson statistics were applied, scaled to a total of 8 million counts for the 60-min scan. This simulation framework is a simplified model that does not account for more complex physical factors, such as positron range, photon non-collinearity, and detector response. To ensure statistical robustness, this entire procedure was repeated 20 times, yielding a set of independent noisy datasets for evaluation.

The performance of the proposed ST-GF method was evaluated against several methods, including BM3D, SDIGF, and two ablative variants of our method: a spatial-only filter (S-GF) and a temporal-only filter (T-GF). All methods first denoise the dynamic sinogram data, after which reconstruction is uniformly performed using the MLEM algorithm for 100 iterations. Its implementation is based on the iterative update rule defined in Equation 2 and was realized using reconstruction code from the KEM Toolbox (https://wanglab.faculty.ucdavis.edu/code). The parameter settings for the compared methods are as follows: BM3D: the parameter σ was set to 15. SDIGF (10): the neighborhood size was set to 10, and the smoothing parameter was set to 0.5.

For quantitative comparison, we employ mean squared error (MSE), Mean Absolute Error (MAE) and bias, variance. The definitions for MSE, Bias, Var, and MAE are provided by:

MSE=jmimage(xj-xjtrue)2jmimage(xjtrue)2,Bias=jmimage(1OiOxji-xjtrue)jmimagexjtrue,Var=1OiOjmimage(xji-1OiOxji)2jmimage(xjtrue)2,MAE=1mimagej=1mimage|xj-xjtrue|

The SSIM compares the similarity between two images and can be expressed as:

SSIM(x,xtrue)=(2μxμxtrue+C1)(2σxxtrue+C2)(μx2+μxtrue2+C1)(σx2+σxtrue2+C2)

where O is the total number of noisy realizations, xjtrue is the j-th element of the true PET activity image, xj is the j-th element of a reconstructed PET image from a single realization, xji is the j-th element of the reconstructed image from the i-th noisy realization, and mimage is the total number of pixels in the image. For SSIM, μx and μxtrue are the local means, σx2 and σxtrue2 are the local variances, and σxxtrue is the local cross-covariance for the reconstructed image x and true image xtrue, respectively. C1 and C2 are stabilization constants to avoid division by zero.

3.2 Implementation details of the proposed method

For the spatial filter construction, the original 24 time frames were first combined into four composite frames (grouped as frames 1–8, 9–16, 17–20, and 21–24) and normalized by their standard deviation. Then, for each composite frame, the k-nearest neighbors (kNN) algorithm was employed to construct a sparse graph by selecting 80 nearest neighbors (k = 80). This value was empirically determined to balance computational complexity and performance. The weights between neighbors were then computed using a Gaussian kernel with a scaling parameter σs = 0.5.

For the temporal filter construction, the sinogram data for each time frame were first pre-smoothed with a 3 × 3 Gaussian filter. A masking technique was then used to focus on high-activity regions, ensuring that the temporal similarity measure is driven by tracer kinetics rather than by low-count background noise. Within these masked regions, a temporal window of size 9 was used to find neighboring frames, and weights were constructed using a Gaussian kernel with a scaling parameter σt = 1.

For all proposed methods (S-GF, T-GF, and ST-GF), an iterative denoising strategy was employed with a stopping threshold of ϵ = 0.01. The weights for the stopping criterion were set to w1 = 0.5 and w2 = 0.5 to give equal importance to both the data and image domains during convergence assessment.

3.3 Results

The results in Figure 3 demonstrate the performance of five denoising methods on dynamic PET sinograms at three different photon count levels: low-count Frame 8 (~61k photons), medium-count Frame 16 (~392k photons), and high-count Frame 24 (~735k photons).

Figure 3
A series of filtered images comparing different denoising methods. Each row includes the ground truth, a noisy sinogram, and results from BM3D, SDIGF, S-GF, T-GF, and ST-GF methods. Mean Squared Error (MSE) values are shown for each method. Row (a) shows high MSE for the noisy image and decreasing MSE values with different filters. Similar patterns follow in rows (b) and (c), with decreasing MSE from noisy to processed images.

Figure 3. Comparative denoising performance for PET sinograms at different statistical levels. The figure showcases three representative time frames, corresponding to low [(a) Frame 8, 61k counts], medium [(b) Frame 16, 392k counts], and high [(c) Frame 24, 735k counts]. For each frame (row), the columns present ground-truth, the noisy and the denoised outputs from the various methods under comparison. The MSE calculated against the ground truth is provided above each corresponding image.

As illustrated in this figure, ST-GF yields the lowest MSE among all methods under the low-count (0.75%) and medium-count (0.27%) conditions. In the high-count frame (Frame 24), however, the BM3D method achieves the lowest MSE of 0.18%, outperforming ST-GF (0.25%). Visually, the result from BM3D exhibits a high degree of smoothness.

Figure 4 illustrates a comparison of the reconstructed images from the denoised sinograms depicted in Figure 3.

Figure 4
Three rows of brain images labeled (a), (b), and (c). Each row shows a progression from a ground truth image, followed by a noise image and five processed images labeled BM3D, SDIGF, S-GF, T-GF, and ST-GF, each with a Mean Squared Error (MSE) percentage. Image quality improves from left to right in each row.

Figure 4. Impact of sinogram denoising on image reconstruction quality. The figure showcases results from three representative time frames, corresponding to low [(a) Frame 8, 61k counts], medium [(b) Frame 16, 392k counts], and high [(c) Frame 24, 735k counts] statistical levels. For each time frame (row), the first two columns display the reference images reconstructed from the ground-truth (noiseless) and the original noisy sinograms, respectively. The subsequent columns present the reconstructed images derived from the outputs of the various sinogram denoising methods shown in Figure 3. The MSE for each reconstructed image, calculated against the ground-truth image, is provided above it.

For low-count Frame 8 (Figure 4a), the ST-GF image shows the highest visual similarity to the ground truth, exhibiting the lowest MSE (4.55%) and the least amount of noise. For lesion visualization under this extremely low signal-to-noise ratio, only ST-GF successfully reconstructs a lesion with a complete and well-defined contour, whereas other methods suffer from blurring, artifacts, or loss of detail.

At the medium-count Frame 16 (Figure 4b), while the quality of all images improves with the increased photon count, the ST-GF result remains superior, achieving the lowest MSE (2.92%), the clearest lesion, and the best texture contrast.

In the high-count Frame 24 (Figure 4c), a key observation is that although BM3D achieved the lowest MSE in the sinogram domain, its corresponding reconstructed image yields a significantly higher MSE (4.88%) than those from the graph filter-based methods. In contrast, the image reconstructed by ST-GF maintains the minimum MSE (2.91%) and the sharpest textures.

Figure 5 compares the average MSE curves over the entire sequence of time frames. Figure 5a illustrates the log(MSE%) for the denoised sinograms, while Figure 5b shows the log (MSE%) for the final reconstructed images. The MSE curves in Figure 5 show that the proposed ST-GF method achieves the lowest MSE at almost all time frames, with its performance being significantly superior to all other compared methods. The advantage of ST-GF is particularly prominent in the early, low-count frames, where its MSE curve is substantially lower than those of the other methods, demonstrating its exceptional robustness in high-noise cases.

Figure 5
Two line graphs compare algorithms' performances using log mean squared error (MSE%) over 25 frame indices. Graph (a) shows MLEM, BM3D, SDIGF, S, T, and ST with varying MSE values. Graph (b) displays similar trends with reduced log MSE%. Each algorithm's performance fluctuates, generally decreasing over time.

Figure 5. MSE comparison curves over all time frames for different methods. (a) Log (MSE%) for denoised sinograms. (b) Log (MSE%) for MLEM reconstructed images. Every point on the curves represents the average of 20 noisy realizations.

Figure 6 plots the Bias-Variance trade-off for the different methods in four regions of interest (ROI) at three distinct time frames (Frame 8, 16, and 24), where the ideal result is located in the bottom-left corner of the graph. The plots demonstrate that the proposed ST-GF method achieves the best bias-variance trade-off across all time frames and all ROIs, as its data point is consistently closest to the origin (the ideal location) in every chart. The advantage of ST-GF is particularly significant in the noisiest scenario, Frame 8. As the photon count increases, the performance of all methods improves, yet ST-GF maintains its comprehensive lead.

Figure 6
Nine line graphs showing Bias squared versus Variance for different conditions: Whole image, Tumor, Graymatter, and Whitematter. Each condition is analyzed in Frames 8, 16, and 24. The graphs compare methods MLEM, BM3D, SDIGF, T, and ST, showing varying performance across conditions and frames.

Figure 6. Bias-Variance trade-off plots for reconstructed images at Frame 8, 16, and 24. Each column corresponds to a different ROI. The ideal performance located in the bottom-left corner (low bias, low variance).

Figure 7 illustrates the effect of the stopping criterion threshold (ϵ) on the final performance of the graph filters, evaluated by (a) Average MSE and (b) Average SSIM. In this analysis, both MSE and SSIM are evaluated. This is necessary because while MSE effectively measures the fidelity of pixel intensities, it is less sensitive to the structural information of an image. The inclusion of SSIM as a complementary metric is crucial as it more accurately quantifies the preservation of anatomical structures and textural details, aspects that are vital for visual assessment in clinical diagnosis. The trends indicate that while performance improves as the threshold becomes stricter, the curves begin to plateau for values less than 0.01. Therefore, 0.01 was selected as the stopping threshold for our final experiments to achieve an optimal balance between performance and computational efficiency.

Figure 7
Two graphs compare S-GF, T-GF, and ST-GF across threshold values. Graph (a) shows average MSE decreasing for all, with ST-GF being the lowest. Graph (b) shows average SSIM increasing, with ST-GF the highest. Threshold values range from zero point five to zero point zero zero zero one.

Figure 7. The effect of the stopping criterion threshold on denoising performance. (a) Average MSE (%) and (b) Average SSIM are plotted against different threshold values for the S-GF, T-GF, and ST-GF methods. A threshold of 0.01 was chosen for the final experiments to balance performance and efficiency.

A quantitative comparison of the MAE, with values averaged over all 24 time frames, is provided in Table 1. The data clearly indicate that our proposed ST-GF method achieves the lowest average MAE in all regions (tumor, gray matter, and white matter), demonstrating its superior overall performance.

Table 1
www.frontiersin.org

Table 1. MAE for different brain regions and reconstruction methods.

3.4 In-vivo results

To further validate the efficacy of ST-GF in a real-world application, experiments were conducted on in-vivo PET data. The data were obtained from the Monash Biomedical Imaging Facility, Australia (28), acquired on a Siemens mMR 3-Tesla scanner. During the scan, a mean dose of ~220 MBq of the 18F-FDG radiotracer was administered to the subject. The total scan duration for the original dataset was 95 min. For this study, we processed the original study's raw data via a rebinning process to generate a dynamic sequence of 26 sinogram frames with a total duration of ~30 min, ensuring our denoising algorithm operated on projection-domain data. Each sinogram was dimensioned at 360 angular projections by 319 radial bins. Attenuation correction was applied to all data with a pseudo-CT. To ensure a fair comparison across all tested methods, a uniform reconstruction pipeline, using the MLEM algorithm with 50 iterations, was subsequently applied to all processed sinograms for final image evaluation.

In the in-vivo experiments, the efficacy of the proposed methods was assessed against several competing techniques. All reconstructions used the MLEM algorithm with 50 iterations. The parameters for each method were adjusted for the characteristics of the real data, as follows: For BM3D, the noise standard deviation parameter σ was set to 9. For SDIGF, a guided filter approach was employed, where the guide image was generated from the sum of all time-frame projections. The window size was selected as 6, and the smoothing parameter was chosen as 0.1. For the proposed graph filter methods (S-GF, T-GF, and ST-GF), the parameters were empirically adjusted for the in-vivo data. For the spatial filter, the frames were divided into three groups to generate composite frames, the kNN neighbor was assigned a value of 150, and the Gaussian kernel σ was fixed at 1. For the temporal filter, the window size was assigned a value of 7, and its Gaussian kernel σ was also fixed at 1. The stopping threshold for the iterative denoising was set to 0.01, consistent with the simulation studies.

Since a “gold standard” ground truth is unavailable for real in-vivo data, we assessed the efficacy of the various methods through a comprehensive approach combining qualitative visual comparison and quantitative contrast-to-noise analysis. We focused on comparing the reconstruction results of representative time frames (Figures 8, 9) and plotted contrast-to-noise curves for regions of interest (ROI) (Figure 10) to quantify the trade-off between signal preservation and noise suppression for each method.

Figure 8
Sequential grid of sinogram images from three frames labeled as Frame 9, Frame 13, and Frame 17. Each frame contains six variations titled Noise sinogram, BM3D, SDIGF, S-GF, T-GF, and ST-GF. The images show different levels and styles of noise reduction or filtering, with adjacent grayscale bars indicating intensity levels.

Figure 8. Comparison of denoised sinograms for the in-vivo dataset. Representative frames for low (Frame 9), medium (Frame 13), and high (Frame 17) count levels are shown.

Figure 9
Series of brain scan images arranged in three rows labeled Frame 9, Frame 13, and Frame 17. Each row contains six columns labeled Noise image, BM3D, SDIGF, S-GF, T-GF, and ST-GF, showing different filtering techniques. The images are color-coded, with scales on the right correlating to numerical values from zero to four thousand or fifteen thousand, indicating intensity levels.

Figure 9. Comparison of reconstructed PET images using denoised sinograms. Representative frames for low (Frame 9), medium (Frame 13), and high (Frame 17) count levels are shown.

Figure 10
(a) Colored heatmap of an anatomical cross-section with six highlighted regions of interest. (b-d) Graphs plotting ROI mean against background noise (normalized STD) with curves for various filtering methods: MLEM, BM3D, SDIGF, S-GF, T-GF, and ST-GF. Each graph shows different noise levels and method performances.

Figure 10. Contrast-noise trade-off curves for various methods under three count levels. (a) The chosen ROI (white box) and background (red box). The curves in (b–d) plot the mean ROI value against the normalized standard deviation (STD) of the background noise for (b) low (Frame 9), (c) medium (Frame 13), and (d) high (Frame 17) count levels, respectively.

The qualitative evaluation begins with the denoised sinograms (Figure 8), with Frames 9, 13, and 17 selected to represent low, medium, and high-count levels, respectively. The original sinograms (first column) exhibit significant granular noise across all frames. The proposed ST-GF method yields the best visual results, producing the smoothest sinograms with the most effective noise suppression at all count levels. In comparison, while the results from BM3D are also very smooth, they appear somewhat blurry visually. The other methods, including SDIGF, S-GF, and T-GF, show noticeable residual noise, particularly in the low-count Frame 9. While their performance improves as the photon count increases, their overall noise control remains inferior to that of ST-GF.

Figure 9 presents the reconstructed PET images. As no ground truth is available for in-vivo data, the evaluation in this section is based on visual quality. In the high-count Frame 17, a noticeable reduction in image noise was observed across all denoising methods, making brain structures clearer than in the original noisy image. However, a careful comparison reveals that the BM3D image suffers from oversmoothing, resulting in a loss of fine structural details. In contrast, all graph filter-based methods (S-GF, T-GF, and ST-GF) achieve better visual results, among which the ST-GF image demonstrates the cleanest background, highest tissue contrast, and the best overall visual quality.

Under the noisier conditions of the medium-count (Frame 13) and low-count (Frame 9), the performance differences become more apparent. In these challenging frames, only the ST-GF method simultaneously suppresses noise effectively while clearly rendering the brain's anatomical structures. The effects of BM3D and SDIGF are limited, with little improvement in image quality. While using the S-GF or T-GF alone provides some improvement, the images still suffer from noticeable noise or blurring. The result from ST-GF is visually superior to those from the S-GF and T-GF methods individually. This observation suggests that the combination of spatial and temporal filtering produces a synergistic effect for achieving the optimal visual outcome.

For a quantitative comparison, contrast-to-noise curves were plotted for the different methods at three representative time frames, as shown in Figure 10. These curves illustrate the trade-off between the average signal from four selected ROI (red boxes, shown in Figure 10a) and the noise STD from the background region (white boxes). An ideal method should produce a curve in the top-left portion of the plot, signifying a higher ROI signal intensity for a given level of background noise.

The plots in Figure 10 demonstrate that the proposed ST-GF method achieves the optimal contrast-noise trade-off in all frames, as its curve is consistently located in the top-left relative to all other methods' curves. Taking the medium-count Frame 13 (Figure 10b) as an example, at the same background noise level of 2.0 × 10−4, the ROI Mean for ST-GF is ~100 units higher than that of the nearest competitor, T-GF. Conversely, for the same ROI Mean of 6,800, the background noise for ST-GF is ~15% lower than that of T-GF. The advantage of ST-GF is especially noticeable in the noisier low- and medium-count frames.

Furthermore, the data show that the performance of the combined ST-GF is superior to that of either single-dimension filtering approach. In contrast, the curves for methods like MLEM and BM3D are positioned in the bottom-right, indicating their lower efficiency in suppressing noise while maintaining signal intensity.

4 Discussion

The observed performance of the BM3D method in this study highlights a potential limitation of traditional denoising approaches when applied to inverse problems like PET reconstruction. In high-count frames, BM3D exhibits excellent denoising performance on the sinograms, achieving an MSE comparable to the proposed ST-GF method (Figure 5a). However, this seemingly strong result in the data domain did not translate to a high-quality final image. On the contrary, the images reconstructed from BM3D-processed sinograms yielded one of the highest MSE values and exhibited significant structural blurring (Figures 4c, 5b). This discrepancy suggests that the objective of a successful denoising algorithm for PET should not be the optimization of metrics in the data domain, but rather fidelity in the final image domain. BM3D's tendency to oversmooth the sinogram, while effective at reducing noise statistically, simultaneously damages the subtle structural information and high-frequency details that are critical for an accurate reconstruction. This fundamentally compromises the integrity of the data-to-image mapping.

The results also demonstrate the importance of addressing both spatial and temporal dimensions simultaneously for effective noise suppression in dynamic PET. The experimental data provide comprehensive validation for this. Across qualitative visual assessments (Figures 4, 9) and all quantitative metrics, including MSE (Figure 5), bias-variance (Figure 6), and contrast-to-noise curves (Figure 10), S-GF or T-GF offers limited improvements. In contrast, the combined ST-GF method demonstrates superior performance in all frames. This indicates that spatial and temporal information are complementary and indispensable in dynamic PET data. The spatial filter is effective for smoothing intra-frame noise by leveraging anatomical consistency, while the temporal filter suppresses abrupt changes between frames by capturing the smooth progression of the tracer over time. Therefore, the strength of the proposed ST-GF framework lies in its unique iterative process: by rebuilding the graph filters at each step from the progressively cleaner signal, it builds a more faithful model of the latent spatio-temporal structure with increasing accuracy. This allows it to achieve a more effective balance between noise reduction and structural preservation.

This unified, model-driven approach represents a distinct pathway in medical image denoising, particularly when contrasted with purely data-driven deep learning methods. The goal of ST-GF is not simply to achieve the highest metrics on a benchmark dataset; rather, its primary advantages are its robustness, adaptability, and interpretability. Its adaptability is rooted in its single-scan nature: by constructing filters for each specific scan, it avoids the “pattern mismatch” problem and potential biases that can arise when a generic model, trained on an external database, is applied to a unique patient case. Furthermore, its interpretability stems from its transparent, model-based process, where each step has a clear mathematical and physical meaning. In clinical practice, the ability to preserve a patient's true anatomical structures with high fidelity may be more valuable than improvements in quantitative metrics alone.

Despite these advantages, the proposed method also has some limitations. Its main limitation is that creating the graph filters requires manual parameter tuning. A key direction for future work is to automate this process using the powerful feature representation of deep neural networks. The main challenge for this is the lack of large, labeled datasets in clinical PET, which makes conventional supervised methods not feasible. Therefore, we propose using self-supervised learning (SSL), an approach that avoids the need for external labels by learning directly from the input data itself. An SSL framework could be used to learn powerful embeddings that capture the data's latent spatio-temporal structure. The similarity between these embeddings would then define the graph adjacency matrix, leading to a highly adaptive, data-driven filter. This would not only eliminate manual parameter tuning but is also expected to greatly improve the filter's reliability and performance.

The simulation experiments in this study were designed to provide a foundational performance validation for the proposed ST-GF framework, and thus we made some methodological simplifications. We recognize that for the ultimate assessment of an algorithm's clinical translation potential, validation on data from Monte Carlo simulation tools, such as GATE, is indispensable. However, for the initial validation of a novel algorithmic framework, employing a more standardized and controlled simulation environment is a common and effective strategy. This approach allows the core mathematical principles of the algorithm to be evaluated separately from complex physical confounders (such as positron range, detector response, and PSF-induced blur), enabling a clear and unambiguous assessment of the denoising mechanism itself. The strong performance of our method on the real in-vivo data, which inherently includes system blur, also validates the effectiveness and robustness of its core spatio-temporal filtering mechanism. Nevertheless, we believe that integrating physical knowledge of the system into the denoising process is key to enhancing algorithmic performance. Future work will be dedicated to explicitly incorporating physical models like the PSF into our ST-GF framework. For instance, the graph could be constructed using a PSF-aware similarity metric to more accurately reflect the underlying structures in the data. We anticipate that this combination of model-driven and data-driven approaches will further improve denoising performance and better preserve fine structural details in the image.

Furthermore, the computational cost of the method is another practical limitation. Our method involves an iterative process where graph construction and filtering are performed at each step, which indeed leads to a high computational cost. We acknowledge this is significantly longer than non-iterative methods like BM3D. On a computing platform with a 12th Gen Intel(R) Core(TM) i5-12,600K (3.69 GHz) CPU, 64GB of RAM, and running MATLAB 2016b, processing the 24-frame simulated dataset took ~85s, while the 26-frame in-vivo dataset required ~102s. However, this computational cost is a direct trade-off for achieving the core advantage of our method: single-scan adaptivity. Unlike deep learning models that require extensive offline training on external data, our method invests its computational resources entirely in an optimal filter for the current scan, which in turn yields superior performance and flexibility. Furthermore, the current CPU implementation is far from optimal. The algorithm, particularly the kNN search, has high potential for parallelization and can be significantly accelerated with GPUs in future work to reduce computation time and enhance clinical utility.

Another area for future improvement is the rule for stopping the iteration. The current rule checks for convergence in both the data (sinogram) and image domains, stopping the process when a simple weighted sum of their changes drops below a set value. While this two-domain rule is stable and practical, this simple combination may not fully capture the complex relationship between the two areas. Future work could therefore explore more advanced, non-linear functions to combine these measurements, which could lead to a more dependable stopping rule that improves the trade-off between reconstruction speed and final image quality.

5 Conclusion

This paper has introduced an advanced, single-scan adaptive spatio-temporal graph filtering (ST-GF) framework that establishes a new paradigm for dynamic PET denoising. The framework's core contribution is a unique iterative process designed to progressively uncover the latent spatio-temporal structure inherent within the data. In this process, the graph filter is not static but is adaptively reconstructed at each iteration from the progressively cleaner signal, allowing it to learn the data's intrinsic structure with increasing fidelity. This structure-discovery process is guided by a second key innovation: a dual-domain stopping criterion that introduces a feedback loop from the image domain, ensuring the entire framework is optimized for superior final image quality.

Extensive simulation and in-vivo results have demonstrated that ST-GF significantly outperforms existing methods, such as BM3D and SDIGF, particularly at low photon counts. Quantitative analyses confirmed its superiority, showing the lowest MAE across all ROIs and an optimal bias-variance trade-off. More importantly, under challenging low-count conditions in both simulation and real data, only ST-GF successfully reconstructed lesions with clear and complete contours, a critical factor for clinical diagnosis. These results consistently validate that the framework's adaptive iterative refinement effectively synergizes spatial and temporal filtering. By achieving this superior performance without the need for external training data, ST-GF establishes itself as a robust, interpretable, and clinically-deployable new paradigm for dynamic PET denoising.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

SG: Conceptualization, Data curation, Formal analysis, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing. XL: Formal analysis, Investigation, Methodology, Validation, Writing – review & editing. SP: Data curation, Writing – review & editing. DZ: Conceptualization, Funding acquisition, Project administration, Resources, Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work is supported by QKHJC[2024]youth190, Guizhou Provincial Basic Research Program (Natural Science) under Grant No. MS [2025]228, Youth Science and Technology Talent Growth Project of Guizhou Provincial Department of Education (Grant Nos. qianjiaoji [2024]74 and qianjiaoji [2024]78), and the National Natural Science Foundation of China (NSFC) under the Grant No. 62363004.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Klyuzhin IS, Cheng JC, Bevington C, Sossi V. Use of a tracer-specific deep artificial neural net to denoise dynamic PET images. IEEE Trans Med Imaging. (2020) 39:366–76. doi: 10.1109/TMI.2019.2927199

PubMed Abstract | Crossref Full Text | Google Scholar

2. Jamet B, Bailly C, Carlier T, Touzeau C, Nanni C, Zamagni E, et al. Interest of PET imaging in multiple myeloma. Front Med. (2019) 6:69. doi: 10.3389/fmed.2019.00069

PubMed Abstract | Crossref Full Text | Google Scholar

3. Gouel P, Decazes P, Vera P, Gardin I, Thureau S, Bohn P. Advances in PET and MRI imaging of tumor hypoxia. Front Med. (2023) 10:1055062. doi: 10.3389/fmed.2023.1055062

PubMed Abstract | Crossref Full Text | Google Scholar

4. Chen S, Liu H, Hu Z, Zhang H, Shi P, Chen Y. Simultaneous reconstruction and segmentation of dynamic PET via low-rank and sparse matrix decomposition. IEEE Trans Biomed Eng. (2015) 62:1784–95. doi: 10.1109/TBME.2015.2404296

PubMed Abstract | Crossref Full Text | Google Scholar

5. Wang G. High temporal-resolution dynamic PET image reconstruction using a new spatiotemporal kernel method. IEEE Trans Med Imaging. (2019) 38:664–74. doi: 10.1109/TMI.2018.2869868

PubMed Abstract | Crossref Full Text | Google Scholar

6. Shepp LA, Vardi Y. Maximum likelihood reconstruction for emission tomography. IEEE Trans Med Imaging. (2007) 1:113–22. doi: 10.1109/TMI.1982.4307558

PubMed Abstract | Crossref Full Text | Google Scholar

7. Lange K, Carson R. EM reconstruction algorithms for emission and transmission tomography. J Comput Assist Tomogr. (1984) 8:306–16.

Google Scholar

8. Liu H, Yang C, Pan N, Song E, Green R. Denoising 3D MR images by the enhanced non-local means filter for Rician noise. Magn Reson Imaging. (2010) 28:1485–96. doi: 10.1016/j.mri.2010.06.023

PubMed Abstract | Crossref Full Text | Google Scholar

9. Peltonen S, Tuna U, Ruotsalainen U. Low count PET sinogram denoising. In: IEEE Nuclear Science Symposium and Medical Imaging Conference Record. Anaheim, CA: IEEE (2012). doi: 10.1109/NSSMIC.2012.6551908

Crossref Full Text | Google Scholar

10. Hashimoto F, Ohba H, Ote K, Tsukada H. Denoising of dynamic sinogram by image guided filtering for positron emission tomography. IEEE Trans Radiat Plasma Med Sci. (2018) 2:541–8. doi: 10.1109/TRPMS.2018.2869936

Crossref Full Text | Google Scholar

11. Yang L, Li Z, Ge R, Zhao J, Si H, Zhang D. Low-dose CT denoising via sinogram inner-structure transformer. IEEE Trans Med Imaging. (2023) 42:910–21. doi: 10.1109/TMI.2022.3219856

PubMed Abstract | Crossref Full Text | Google Scholar

12. Kruzhilov I, Kudin S, Vetoshkin L, Sokolova E, Kokh V. Whole-body PET image denoising for reduced acquisition time. Front Med. (2024) 11:1415058. doi: 10.3389/fmed.2024.1415058

PubMed Abstract | Crossref Full Text | Google Scholar

13. Jaudet C, Weyts K, Lechervy A, Batalla A, Bardet S, Corroyer-Dulmont A. The impact of artificial intelligence CNN based denoising on FDG PET radiomics. Front Oncol. (2021) 11:692973. doi: 10.3389/fonc.2021.692973

PubMed Abstract | Crossref Full Text | Google Scholar

14. Shuman DI, Narang SK, Frossard P, Ortega A, Vandergheynst P. The emerging field of signal processing on graphs: extending high-dimensional data analysis to networks and other irregular domains. IEEE Signal Process Mag. (2013) 30:83–98. doi: 10.1109/MSP.2012.2235192

Crossref Full Text | Google Scholar

15. Leus G, Marques AG, Moura JM, Ortega A, Shuman DI. Graph signal processing: history, development, impact, and outlook. IEEE Signal Process Mag. (2023) 40:49–60. doi: 10.1109/MSP.2023.3262906

Crossref Full Text | Google Scholar

16. Guo S, Sheng Y, Chai L, Zhang J, PET. Image reconstruction with Kernel and Kernel space composite regularizer. IEEE Trans Med Imaging. (2023) 42:1786–98. doi: 10.1109/TMI.2023.3239929

PubMed Abstract | Crossref Full Text | Google Scholar

17. Guo S, Sheng Y, Xiong D, Zhang J. PET image representation and reconstruction based on graph filter. IEEE Trans Comput Imaging. (2023) 9:808–18. doi: 10.1109/TCI.2023.3308388

Crossref Full Text | Google Scholar

18. Guo S, Sheng Y, Chai L, Zhang J. Graph filtering approach to PET image denoising. In: 2019 1st International Conference on Industrial Artificial Intelligence (IAI). Shenyang: IEEE (2019). p. 1–6. doi: 10.1109/ICIAI.2019.8850802

PubMed Abstract | Crossref Full Text | Google Scholar

19. Guo S, Sheng Y, Chai L, Zhang J. Kernel graph filtering–a new method for dynamic sinogram denoising. PLoS ONE. (2021) 16:e0260374. doi: 10.1371/journal.pone.0260374

PubMed Abstract | Crossref Full Text | Google Scholar

20. Belcari N, Bisogni M, Del Guerra A. Positron emission tomography: its 65 years and beyond. La Riv Nuovo Cimento. (2023) 46:693–785. doi: 10.1007/s40766-024-00050-3

Crossref Full Text | Google Scholar

21. Pain CD, Egan GF, Chen Z. Deep learning-based image reconstruction and post-processing methods in positron emission tomography for low-dose imaging and resolution enhancement. Eur J Nucl Med Mol Imaging. (2022) 49:3098–118. doi: 10.1007/s00259-022-05746-4

PubMed Abstract | Crossref Full Text | Google Scholar

22. Qi J, Leahy RM. Iterative reconstruction techniques in emission computed tomography. Phys Med Biol. (2006) 51:R541–78. doi: 10.1088/0031-9155/51/15/R01

PubMed Abstract | Crossref Full Text | Google Scholar

23. Isufi E, Gama F, Shuman DI, Segarra S. Graph filters for signal processing and machine learning on graphs. IEEE Trans Signal Process. (2024) 72:4745–81. doi: 10.1109/TSP.2024.3349788

Crossref Full Text | Google Scholar

24. Kheradmand A, Milanfar P. A general framework for regularized, similarity-based image restoration. IEEE Trans Image Process. (2014) 23:5136–51. doi: 10.1109/TIP.2014.2362059

PubMed Abstract | Crossref Full Text | Google Scholar

25. Wang G, Qi J. PET image reconstruction using Kernel method. IEEE Trans Med Imaging. (2015) 34:61–71. doi: 10.1109/TMI.2014.2343916

PubMed Abstract | Crossref Full Text | Google Scholar

26. Feng D, Wong KP, Wu CM, Siu WC, A. technique for extracting physiological parameters and the required input function simultaneously from PET image measurements: theory and simulation study. IEEE Trans Inf Technol Biomed. (1997) 1:243–54. doi: 10.1109/4233.681168

PubMed Abstract | Crossref Full Text | Google Scholar

27. Spence AM, Muzi M, Graham MM, O'Sullivan F, Krohn KA, Link JM, et al. Glucose metabolism in human malignant gliomas measured quantitatively with PET, l-[C-ll] glucose and FDG: Analysis of the FDG lumped constant. J Nuclear Med. (1998) 39:440–8.

Google Scholar

28. Jamadar SD, Ward PG, Liang EX, Orchard ER, Chen Z, Egan GF. Metabolic and hemodynamic resting-state connectivity of the human brain: a high-temporal resolution simultaneous BOLD-fMRI and FDG-fPET multimodality study. Cereb Cortex. (2021) 31:2855–67. doi: 10.1093/cercor/bhaa393

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: dynamic positron emission tomography, sinogram denoising, graph signal processing, spatio-temporal filtering, graph filter

Citation: Guo S, Li X, Pan S and Zhang D (2025) Single-scan adaptive graph filtering for dynamic PET denoising by exploring intrinsic spatio-temporal structure. Front. Med. 12:1659122. doi: 10.3389/fmed.2025.1659122

Received: 03 July 2025; Accepted: 11 August 2025;
Published: 29 August 2025.

Edited by:

Zekuan Yu, Fudan University, China

Reviewed by:

Daniele Panetta, National Research Council (CNR), Italy
Gengsheng Xie, Jiangxi Normal University, China

Copyright © 2025 Guo, Li, Pan and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Dan Zhang, Y3N6aGRAbWFpbC5ndWZlLmVkdS5jbg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.