Abstract
Spiking Neural Networks (SNNs) offer a paradigm of energy-efficient, event-driven computation that is well-suited for processing asynchronous sensory streams. However, training deep SNNs robustly in an online and continual manner remains a formidable challenge. Standard Backpropagation-through-Time (BPTT) suffers from a prohibitive memory bottleneck due to the storage of temporal histories, while local plasticity rules often fail to balance the trade-off between rapid acquisition of new information and the retention of old knowledge (the stability-plasticity dilemma). Motivated by the tripartite synapse in biological systems, where astrocytes regulate synaptic efficacy over slow timescales, we propose Astrocyte-Gated Multi-Timescale Plasticity (AGMP). AGMP is a scalable, online learning framework that augments eligibility traces with a broadcast teaching signal and a novel astrocyte-mediated gating mechanism. This slow astrocytic variable integrates neuronal activity to dynamically modulate plasticity, suppressing updates in stable regimes while enabling adaptation during distribution shifts. We evaluate AGMP on a comprehensive suite of neuromorphic benchmarks, including N-Caltech101, DVS128 Gesture, and Spiking Heidelberg Digits (SHD). Experimental results demonstrate that AGMP achieves accuracy competitive with offline BPTT while maintaining constant temporal memory complexity. Furthermore, in rigorous Class-Incremental Continual Learning scenarios (e.g., Split CIFAR-100), AGMP significantly mitigates catastrophic forgetting without requiring replay buffers, outperforming state-of-the-art online learning rules. This work provides a biologically grounded, hardware-friendly path toward autonomous learning agents capable of lifelong adaptation.
1 Introduction
Spiking Neural Networks (SNNs) have emerged as a promising paradigm for energy-efficient machine intelligence, offering a computational substrate that closely mimics the sparse, event-driven dynamics of biological brains (Maass, 1997; Roy et al., 2019). Unlike traditional Artificial Neural Networks (ANNs) that process continuous-valued activations synchronously, SNNs communicate via binary spikes asynchronously. This temporal sparsity is particularly advantageous when processing dynamic sensory streams from neuromorphic hardware, such as Dynamic Vision Sensors (DVS) and silicon cochleas, where information is encoded in the precise timing of events rather than static frames (Gallego et al., 2022). As the demand for edge computing and low-latency deployment grows, developing deep SNNs that can learn robustly and efficiently on neuromorphic chips has become a central objective in the field.
However, training deep SNNs to high accuracy remains a formidable challenge due to the non-differentiable nature of the spike generation function. The current state-of-the-art approach, Backpropagation-through-Time (BPTT) with surrogate gradients, circumvents this issue by smoothing the derivative of the Heaviside step function (Wu et al., 2018; Neftci et al., 2019). While BPTT has enabled SNNs to achieve performance competitive with ANNs on complex recognition tasks, it is fundamentally an offline and memory-intensive algorithm. BPTT requires unrolling the network over the entire simulation time window T and storing the intermediate membrane potentials of all neurons to compute gradients. This results in a memory complexity that scales linearly with time (), creating a prohibitive “memory wall” for processing long temporal sequences or for deploying learning algorithms directly on resource-constrained edge devices (Zenke et al., 2021). Furthermore, BPTT assumes that the entire dataset is available for repeated offline epochs, a condition that holds for static benchmarks but fails in real-world scenarios where data arrives in a continuous, non-stationary stream.
To address the limitations of offline BPTT, significant research effort has shifted toward synapse-local plasticity rules inspired by biological mechanisms, such as Spike-Timing-Dependent Plasticity (STDP) (Dan and Poo, 2004). While classical pair-based STDP is computationally efficient and locally implementable, it often struggles to solve complex credit assignment problems in deep, hierarchical networks. Recent advances in “three-factor” learning rules, which combine local presynaptic and postsynaptic activities with a global modulatory signal (e.g., error or reward), have begun to bridge the gap between biological plausibility and deep learning performance (Frémaux and Gerstner, 2016). Notably, the eligibility propagation (e-prop) framework (Bellec et al., 2020) provides a mathematical justification for online learning, factoring the gradient into a fast, synapse-local eligibility trace and a broadcast learning signal. These methods allow weights to be updated at every time step without storing the full history of neural states, theoretically enabling online learning with constant memory complexity ().
Despite these advances in online learning, a critical unresolved issue is Continual Learning (CL). In realistic deployment scenarios, an intelligent agent must learn a sequence of tasks or adapt to changing data distributions over time. When standard plasticity rules (including three-factor methods) are applied to sequential tasks, they suffer from “catastrophic forgetting, where the adaptation to new information rapidly overwrites previously acquired knowledge” (Kirkpatrick et al., 2017; Parisi et al., 2019). In the ANN literature, this stability-plasticity dilemma is often addressed using replay buffers or regularization techniques that require computing the Fisher Information Matrix (e.g., Elastic Weight Consolidation). However, these methods are computationally heavy and often incompatible with the strict locality and online constraints of neuromorphic hardware. Achieving robust continual learning in SNNs requires a mechanism that can regulate plasticity dynamically, stabilizing important synapses while allowing others to adapt, without relying on external memory buffers or offline batch processing.
Theoretical neuroscience suggests that the stability of biological memory arises from the interaction of multiple processes operating across distinct timescales (Benna and Fusi, 2016; Zhang et al., 2022). While neurons and synapses operate on the millisecond scale, non-neuronal glial cells, particularly astrocytes, regulate synaptic function on timescales of seconds to minutes. The concept of the “Tripartite Synapse” posits that astrocytes are active partners in computation, integrating neuronal activity and releasing gliotransmitters that gate synaptic plasticity (Perea et al., 2009; Stogsdill and Eroglu, 2017). This slow, homeostatic regulation serves as a natural mechanism for metaplasticity, determining when and how much a synapse should change based on the broader context of neuronal activity.
Motivated by these biological principles, we propose Astrocyte-Gated Multi-Timescale Plasticity (AGMP), a novel online learning framework for deep SNNs. AGMP extends the three-factor eligibility trace paradigm by introducing a fourth factor: an astrocyte-like slow state variable that integrates local neuronal activity. This slow variable acts as a dynamic gate for synaptic updates, effectively modulating the learning rate based on the historical activity context of the neuron. By coupling fast eligibility traces (for temporal credit assignment) with slow astrocytic gating (for stability) and a broadcast error signal (for task performance), AGMP enables deep SNNs to learn continuously from streaming data. We demonstrate that this multi-timescale approach not only matches the accuracy of offline BPTT on event-based neuromorphic benchmarks, such as DVS128 Gesture and SHD, but also significantly mitigates catastrophic forgetting in Class-Incremental and Task-Incremental learning scenarios. This work provides a scalable, biologically interpretable path toward robust online learning in neuromorphic systems, ensuring more trustworthy computing in decision-making (Song and Zhang, 2025).
This paper makes the following contributions:
-
A structured four-factor learning mechanism. We introduce an astrocyte-like slow state that gates eligibility-trace learning driven by a broadcast modulatory teaching signal, yielding an online four-factor plasticity rule: eligibility × modulation × astrocytic gate × stabilization.
-
Deep and online learning without BPTT histories. AGMP maintains only synapse-local eligibility traces and neuron-local astrocyte states, avoiding the need to store T-step histories typical of BPTT and enabling scalable online updates in deep CNN-SNN and spiking ResNet architectures.
-
Continual learning evaluation under Task-IL and Class-IL. We evaluate AGMP under both task-incremental (multi-head) and class-incremental (single-head) protocols (van de Ven and Tolias, 2019), highlighting the role of astrocytic gating in reducing interference under the more challenging Class-IL setting.
-
Energy-aware reporting and mechanistic ablations. In addition to accuracy, we report spike counts and synaptic operations (SynOps) as energy proxies, and we include ablations (gate, eligibility, homeostasis) and time-scale sensitivity sweeps to isolate the contributions of each component.
2 Related work
2.1 Deep SNN training with surrogate gradients
Training SNNs has historically been impeded by the non-differentiable nature of discrete spike events. Early approaches largely relied on ANN-to-SNN conversion (Diehl et al., 2015; Sengupta et al., 2019; Zhang et al., 2024), where a rate-coded SNN approximates a pre-trained ANN. While conversion methods yield high accuracy on static image datasets, they typically suffer from high inference latency and fail to capture the rich temporal dynamics required for event-based sensory processing.
To enable direct training in the temporal domain, the concept of Surrogate Gradients (SG) was introduced (Neftci et al., 2019; Zenke and Ganguli, 2018; Wu et al., 2018). By replacing the derivative of the Heaviside step function with a smooth auxiliary function (e.g., sigmoid or arctangent) during the backward pass, SG allows SNNs to be optimized via Backpropagation-through-Time (BPTT). This paradigm has established the current state-of-the-art for deep SNNs on complex neuromorphic benchmarks (Shrestha and Orchard, 2018; Yin et al., 2020; Wu et al., 2025). However, BPTT is inherently non-local in time; it requires unfolding the network graph and storing synaptic and membrane states for the entire duration of the input sequence. This results in a memory complexity of , rendering the training of long sequences on memory-constrained edge devices computationally prohibitive (Zenke et al., 2021).
2.2 Online learning and eligibility traces
To overcome the “memory wall” of BPTT, recent research has pivoted toward forward-mode learning rules that rely on local approximations of the gradient. Inspired by the biological three-factor learning rule, which involves presynaptic activity, postsynaptic state, and a global modulator (Frémaux and Gerstner, 2016; Porr and Wörgötter, 2007), researchers have formalized these dynamics into efficient algorithms.
The e-prop algorithm (Bellec et al., 2020) represents a significant breakthrough, mathematically decomposing the gradient into a synapse-local eligibility trace and a broadcast learning signal. This allows weights to be updated online at every time step without storing future gradients. Similar approaches, such as Online Spatio-Temporal Learning (OSTL) (Bohnstingl et al., 2023) and Super-Spike (Zenke and Ganguli, 2018), share this philosophy of separating temporal credit assignment (trace) from spatial error assignment. While these methods achieve temporal memory complexity, they primarily address the online aspect of learning rather than the continual aspect. When exposed to non-stationary data streams without explicit regularization, these plasticity rules are prone to rapid weight overwriting, leading to performance degradation on previously learned tasks.
2.3 Continual learning and stability-plasticity
Continual Learning (CL) deals with learning from a stream of data where the distribution changes over time, posing the “stability-plasticity dilemma” (Parisi et al., 2019). In the context of ANNs, prominent strategies include regularization-based methods like Elastic Weight Consolidation (EWC) (Kirkpatrick et al., 2017) and Synaptic Intelligence (SI) (Zenke et al., 2017), which penalize changes to parameters critical for past tasks. However, these methods typically require calculating the Fisher Information Matrix or path integrals, which are computationally expensive to estimate in a purely online, streaming setting.
In SNNs, CL is often approached via architecture expansion (Rusu et al., 2022) or replay buffers (Rebuffi et al., 2017), but these solutions entail growing memory costs that negate the efficiency benefits of neuromorphic hardware. Recent efforts have explored metaplasticity—the plasticity of plasticity itself. For instance, models incorporating hidden synaptic states (Benna and Fusi, 2016) or binary synapses with probabilistic transitions (Fusi et al., 2005) can extend memory retention. Nevertheless, integrating these theoretical mechanisms into deep, supervised SNNs capable of processing high-dimensional inputs (e.g., DVS gestures) remains an open challenge. Our work addresses this by implementing metaplasticity not through hidden weights, but through an explicit, biologically grounded gating variable.
2.4 Astrocytes in neuromorphic computing
Biological synapses are not merely bipartite connections between neurons; they are tripartite structures sheathed by astrocytes (Araque et al., 1999; Perea et al., 2009). Neuroscience evidence suggests that astrocytes integrate neuronal activity over slow timescales (seconds) and release gliotransmitters (e.g., glutamate, D-serine) that regulate the threshold for synaptic plasticity (Bernardinelli et al., 2014; Stogsdill and Eroglu, 2017).
Computational models of neuron-glia interactions have traditionally focused on unsupervised tasks, such as self-repair (Wade et al., 2012), synchronization (Chen and Li, 2024), or unsupervised feature clustering (Shen et al., 2023). More recently, researchers have begun to explore astrocytes for supervised learning. For example, Rastogi et al. (2021) proposed astrocyte-modulated STDP to stabilize learning rates, and Reichenbach et al. Reichenbach et al. (2010) demonstrated that glial-like regulation can extend memory retention in simple networks. However, these works often rely on shallow architectures or simplified tasks. AGMP distinguishes itself by mathematically formulating the astrocyte as a gating factor for SG within an eligibility trace framework, enabling scalable, supervised continual learning in deep convolutional and recurrent SNNs.
3 Proposed method
In this section, we present the theoretical formulation of Astrocyte-Gated Multi-Timescale Plasticity (AGMP). We begin by defining the underlying spiking neuron models and the problem of gradient-based learning in discrete time. We then derive the online learning rule using eligibility traces and introduce the novel astrocytic gating mechanism designed to regulate plasticity dynamics for continual learning stability.
3.1 Spiking neuron dynamics
We adopt the Leaky Integrate-and-Fire (LIF) model and its adaptive variant (ALIF) as the computational units. These models capture the essential temporal dynamics of biological neurons while remaining computationally tractable for deep network simulation.
Consider a deep SNN with layers ℓ = 1, …, L. The membrane potential of neuron i in layer ℓ at discrete time step t evolves according to:
where α = exp(−Δt/τm) represents the leakage factor with membrane time constant τm, is the synaptic weight from presynaptic neuron j, is a bias term, and denotes the input spike. The output spike is generated via a deterministic threshold mechanism:
where H(·) is the Heaviside step function. Following a spike, the membrane potential is reset by subtracting the threshold value (soft reset). To enhance temporal processing capabilities, particularly for datasets like SHD, we employ Adaptive LIF (ALIF) neurons. In ALIF, the threshold is dynamic:
where ʒ0 is the baseline threshold, ρ determines the decay of the adaptation, and β controls the magnitude of threshold increase post-spike. This adaptation effectively models firing rate adaptation, providing the network with a longer working memory.
3.2 Gradient descent with eligibility traces
The core challenge in training SNNs is the non-differentiable nature of the discrete spike function S(·). We utilize the surrogate gradient method (Neftci et al., 2019), where the derivative is approximated during learning by a smooth function ψ(·), such as a piecewise linear or sigmoid function:
Standard BPTT computes the exact gradient of the loss by unrolling the network over time. However, this violates the requirement for online learning. To achieve locality in time, we adopt the factorization proposed in the e-prop framework (Bellec et al., 2020). The gradient of the loss with respect to a weight wij at time t is approximated as the product of a global learning signal and a local eligibility trace:
The eligibility trace maintains a fading memory of the correlation between presynaptic input and postsynaptic state. For LIF neurons, this is derived recursively:
where λe relates to the membrane time constant. This variable is purely local to the synapse (i, j) and can be computed forward in time. For the learning signal, we employ a top-down error broadcast. In the output layer, . For hidden layers, exact backpropagation of error requires symmetric weight transport (WT), which is biologically implausible. We instead use direct error broadcast, projecting the global error vector δ[t] to hidden neurons via fixed random matrices Bℓ:
This signal serves as the “third factor” in the plasticity rule, informing the neuron of its contribution to the global objective. To ensure the error signal remains within a range that the astrocyte-gated update can effectively regulate, the elements of the fixed random matrices Bℓ are initialized using a variance-preserving heuristic (scaled by ), preventing the gate from struggling to converge due to initially exploding error magnitudes.
3.3 Astrocyte-gated multi-timescale plasticity (AGMP)
While the three-factor rule described above enables online learning, it treats all updates with equal plasticity potential, leading to catastrophic forgetting in non-stationary environments. We propose AGMP as shown in Figure 1, which introduces a fourth factor: a slow, astrocyte-mediated gating variable.
Figure 1

Schematic of Astrocyte-Gated Multi-Timescale Plasticity (AGMP). The framework mimics the tripartite synapse. Neuronal dynamics (blue) operate on a fast timescale (τm). Synaptic eligibility traces (light blue) capture millisecond correlations (τe). The astrocyte (orange) integrates local activity over a slow timescale (τa) to compute a gating factor gi. This gate modulates the update driven by the broadcast error signal (red), enabling the network to balance plasticity (during new tasks) and stability (during consolidation).
3.3.1 Biological motivation and dynamics
In the brain, astrocytes ensheath synapses and regulate their efficacy. Unlike neurons, which operate on the millisecond timescale, astrocytes integrate activity over seconds to minutes. We hypothesize that this slow integration provides a robust estimate of the contextual stability or metabolic cost of a circuit, which can be used to modulate plasticity rates dynamically. We associate a scalar astrocyte state with each neuron i, which evolves on a slow timescale τa ≫ τm, τe:
where λa = exp(−Δt/τa). The drive function represents the local activity load:
where ηu, ηs, and ηo are positive scaling coefficients that control the relative contribution of membrane potential, synaptic input, and output spikes, respectively. This formulation implies that the astrocyte integrates presynaptic drive, postsynaptic potential, and output spiking. A high value of ai indicates a neuron that has been consistently active or highly stimulated over the recent past.
3.3.2 Context-aware gating and update rule
To convert the astrocyte state into a plasticity modulator, we employ a normalization and gating function (Figure 2). At t = 0, the astrocyte state is initialized to . We then compute the layer-wise moving statistics (mean and variance ) of the astrocyte states. The normalized state is . The gating factor is defined as , where σ(·) is the sigmoid function. Note that is indexed by the postsynaptic neuron i, meaning it acts as a postsynaptic-shared modulator for all incoming synapses to neuron i, biologically consistent with astrocytic domains that regulate local volumes. Functionally, this acts as a dynamic learning rate similar to Adam or RMSProp, but driven by local metabolic activity rather than gradient statistics.
Figure 2

Astrocyte Gating Mechanism. The local activity inputs (membrane potential, synaptic input, and output spikes) drive a slow state integrator. This slow state ai is normalized and passed through a sigmoid function to produce the gating factor gi∈[0, 1]. This gate essentially determines the ”write permission” for the synaptic weights connected to neuron i.
As given in Figure 3, combining the eligibility trace, the broadcast error, and the astrocytic gate, the final synaptic weight update at time t is:
While the integration of astrocyte states and gating introduces additional operations, these are neuron-wise computations scaling as . Given that deep networks are typically densely connected where synapses S≫N, the relative computational overhead (FLOPs) of the gate is asymptotically negligible compared to the synaptic operations.
Figure 3

Online Processing Loop. Detailed data flow for a single time step t. The diagram illustrates how local activity ϕi drives the slow astrocyte state ai to generate gate gi. Simultaneously, spikes drive the fast eligibility trace eij. These factors combine with the broadcast error Mi to update weights Δwij purely online, without storing history.
This rule is strictly local in space (except for the broadcast error) and local in time. It requires no storage of history beyond the current simulation step, satisfying the memory constraint. Additionally, to counteract firing rate drift, we implement a lightweight homeostatic mechanism on the bias term , ensuring neurons remain in a sensitive dynamic range.
In terms of computational complexity, AGMP offers a distinct advantage. For a network with N neurons and S synapses trained over T time steps, AGMP requires space complexity, which is constant in time, whereas BPTT requires . Both methods incur compute operations, but AGMP's constant memory footprint enables the training of deeper networks on longer sequences without memory overflow.
4 Experimental results
In this section, we provide a comprehensive evaluation of the proposed AGMP framework. To rigorously validate the effectiveness of our approach, we expand our evaluation across a wide range of datasets and compare against an extensive set of baselines, including both offline gradient-based methods and state-of-the-art online learning rules.
4.1 Experimental setup
To assess performance across varying spatiotemporal complexities, we utilized a diverse suite of neuromorphic and audio datasets. These include N-MNIST (Orchard et al., 2015), a neuromorphic version of MNIST serving as a standard benchmark; DVS128 Gesture (Amir et al., 2017), which requires temporal feature extraction from dynamic hand movements; and N-Caltech101 (Orchard et al., 2015), a challenging event-based object recognition dataset characterized by variable object scales. For high-precision temporal processing, we employed Spiking Heidelberg Digits (SHD) (Cramer et al., 2020) and the large-scale Google Speech Commands (GSC) (Warden, 2018).
Our comparison baselines encompass three categories: Offline Supervised Learning methods including STBP (Wu et al., 2018) and Diet-SNN (Rathi and Roy, 2023); Online/Local Learning rules such as e-prop (Bellec et al., 2020), OSTL (Bohnstingl et al., 2023), and the recent OTTT (Xiao et al., 2022a); and Continual Learning strategies including EWC (Kirkpatrick et al., 2017), SI (Zenke et al., 2017), LwF (Li and Hoiem, 2018), and the replay-based GEM (Lopez-Paz and Ranzato, 2017). We maintained a consistent architecture strategy, utilizing a 7-layer VGG-SNN for N-Caltech101 and DVS Gesture, and a ResNet-18 SNN for CIFAR-based tasks. For audio tasks, a Recurrent SNN (RSNN) with 256 ALIF units was employed. All results are averaged over 5 random seeds.
4.2 Performance on standard benchmarks
Table 1 presents the comparison results on both static and temporal neuromorphic classification tasks. As visualized in Figure 4 and detailed in the table, AGMP consistently outperforms existing online learning rules (e.g., e-prop, OSTL) and demonstrates competitiveness against offline BPTT methods.
Table 1
| Method | Mode | N-MNIST | DVS gesture | N-Caltech101 | CIFAR10-DVS | SHD |
|---|---|---|---|---|---|---|
| Offline / Backpropagation-Through-Time (BPTT) | ||||||
| STBP (Wu et al., 2018) | Off | 99.23 | 96.87 | 78.01 | 60.30 | – |
| Diet-SNN (Rathi and Roy, 2023) | Off | 99.20 | 95.90 | 74.50 | – | – |
| SMA-SNN (Shan et al., 2025) | Off | – | – | 84.60 | 84.00 | - |
| Online / local learning rules | ||||||
| e-prop (Bellec et al., 2020) | On | 97.80 | 93.10 | 66.40 | – | 76.20 |
| OTTT (Xiao et al., 2022b) | On | – | 96.88 | – | 76.27 | 77.33 |
| OSTL (Bohnstingl et al., 2023) | On | 98.10 | 91.50 | 65.80 | – | 71.40 |
| BrainScale (ES-D-RTRL) (Wang et al., 2024) | On | 97.45 | 97.29 | – | – | 97.29 |
| Traces Prop. (Pes et al., 2026) | On | 98.45 | 97.33 | – | – | 97.33 |
| S-TLLR (Apolinario and Roy, 2024) | On | – | 97.72 | 74.40 | 75.60 | 78.23 |
| RPLIF (Li et al., 2025) | On | – | 97.22 | 83.35 | 82.40 | – |
| AGMP (ours) | On | 99.10 | 95.40 | 73.80 | 78.50 | 80.10 |
Test accuracy (%) comparison on neuromorphic and audio datasets.
“Mode” indicates Offline (Off) or Online (On) weight updates. “–” indicates the metric was not reported. Best results for each category are in bold.
Figure 4

AGMP Performance on Standard Benchmarks. (a) Test Accuracy on DVS Gesture, CIFAR10-DVS, and SHD compared to BPTT+SG (offline) and e-prop (online). AGMP matches BPTT accuracy within error margins. (b, c) Energy Efficiency: AGMP significantly reduces Spike Counts and Synaptic Operations (SynOps), demonstrating higher sparsity. (d) Stability Indicators: The gate distribution (left) shows selective plasticity, while firing rate trajectories (right) confirm that AGMP prevents the drift observed in unregulated online learning.
Specifically, on the N-Caltech101 dataset, which is characterized by significant noise and class imbalance, AGMP achieves a test accuracy of 73.80%. This represents a substantial improvement over standard eligibility trace methods like e-prop (66.40%) and OSTL (65.80%), indicating that the astrocytic gating mechanism effectively filters irrelevant noisy updates.
On the temporal processing benchmark SHD, AGMP reaches 80.10%, outperforming the recent state-of-the-art online algorithm OTTT (77.33%) and S-TLLR (78.23%). This result confirms that integrating multi-timescale plasticity allows the network to capture long-term temporal dependencies more effectively than fixed-time-constant traces. Furthermore, on CIFAR10-DVS, AGMP achieves 78.50%, surpassing classic online approaches and approaching the performance of complex offline models. Notably, AGMP bridges the accuracy gap between online and offline learning while maintaining the strict locality required for neuromorphic hardware.
4.3 Continual learning performance
4.3.1 Continual learning performance
We evaluate CL performance on two distinct setups: Split N-MNIST (5 tasks, 2 classes each) and the significantly more challenging Split CIFAR-100 (10 tasks, 10 classes each) under the Class-Incremental (Class-IL) setting. Figure 5 and Table 2 compares AGMP against regularization-based, distillation-based, and replay-based methods.
Figure 5

Continual Learning Results. (a) Average Accuracy (AT) and (b) Forgetting (FT) on Permuted MNIST and Split CIFAR-100 benchmarks. AGMP (green) maintains higher accuracy and significantly lower forgetting compared to EWC (orange) and standard online learning (blue), particularly in the later tasks. The astrocytic gate suppresses disruptive updates to consolidated weights.
Table 2
| Method | Type | Split N-MNIST | Split CIFAR-100 (10 Tasks) | ||
|---|---|---|---|---|---|
| A 5(↑) | F 5(↓) | A 10(↑) | F 10(↓) | ||
| EWC (Kirkpatrick et al., 2017) | Regularization | 72.40 | 24.50 | 17.25 | 61.20 |
| SI (Zenke et al., 2017) | Regularization | 71.80 | 25.10 | 17.26 | 62.50 |
| LwF (Li and Hoiem, 2018) | Distillation | 68.50 | 28.40 | 24.20 | 65.80 |
| GEM* (Lopez-Paz and Ranzato, 2017) | Replay | 88.50 | 8.20 | 45.20 | 35.10 |
| DSD-SNN (Han et al., 2023b) | Structural | 97.06 | – | 60.47 | – |
| SOR-SNN (Han et al., 2023a) | Self-Org. | – | - | 80.12 | – |
| SESLR (Lin et al., 2025) | Latent Replay | 95.38 | – | – | – |
| AGMP (ours) | Gating (Metaplasticity) | 82.60 | 15.40 | 31.40 | 55.70 |
Continual learning average accuracy (AT, %) and forgetting (FT, %) on class-incremental tasks.
Comparison includes regularization, replay, and dynamic structure methods. *Methods using explicit replay buffers. Best results for each category are in bold.
AGMP demonstrates remarkable robustness against catastrophic forgetting. On Split N-MNIST, it achieves an average accuracy (A5) of 82.60%, significantly outperforming regularization methods such as EWC (72.40%) and SI (71.80%). Crucially, the forgetting rate (F5) of AGMP is limited to 15.40%, whereas EWC suffers from a 24.50% drop. Unlike EWC, which estimates synaptic importance via a static Fisher Information Matrix only at task boundaries, AGMP's astrocytic gate continuously evolves, providing a dynamic and context-aware measure of synaptic stability.
On the large-scale Split CIFAR-100 benchmark, the advantages of AGMP are even more pronounced. It achieves a final accuracy of 31.40%, nearly doubling the performance of EWC (17.25%) and SI (17.26%). While the replay-based method GEM achieves higher accuracy (45.20%) by explicitly storing raw input samples, AGMP offers a compelling trade-off: it significantly outperforms non-replay baselines and bridges the gap toward replay methods without the privacy concerns and memory overhead associated with data buffering.
4.4 Efficiency and ablation analysis
In terms of efficiency, AGMP reduces SynOps by ~40% compared to standard BPTT and ~15% compared to OTTT (Figure 4c), as homeostatic regulation encourages sparse representations. Crucially, regarding scalability, AGMP maintains a constant memory footprint (~0.9GB for ResNet-18), whereas BPTT's memory usage grows linearly, often causing Out-Of-Memory errors on long sequences (Figure 6).
Figure 6

Ablation and sensitivity studies. (Top Row) Impact of removing specific components (Gate, Eligibility Trace, Homeostasis) on Accuracy, Forgetting, and Spike Count. The full AGMP model yields the best balance. (Bottom Left) Time-scale sweep heatmap: Forgetting is minimized when the astrocyte timescale τa is significantly larger than the eligibility timescale τe (τe≪τa). (Bottom Right) Distribution of gate values and firing rates.
We conducted extensive ablation studies to decouple the contributions of each component (Table 3). Removing the astrocytic gate caused a negligible drop on stationary tasks but a catastrophic 31.3% performance collapse on Split N-MNIST, confirming the gate's specific role in stability. Conversely, removing eligibility traces caused a massive 21.8% drop on temporal tasks (SHD), validating the necessity of traces for temporal credit assignment. Sensitivity analysis revealed that optimal stability is achieved when the astrocyte timescale τa is 50 × to 500 × larger than the eligibility timescale τe, supporting the multi-timescale hypothesis. However, an upper bound exists for τa; if the timescale becomes too large (or effectively infinite) relative to the task duration, the astrocyte gate becomes too stiff to adapt, preventing the learning of new information (under-plasticity).
Table 3
| Configuration | SHD (temporal) | Split N-MNIST (CL) | ||
|---|---|---|---|---|
| Acc (%) | Δ | A 5 (%) | Δ | |
| Full AGMP | 80.1 | – | 82.6 | – |
| w/o Astrocytic gate | 79.5 | –0.6 | 51.3 | –31.3 |
| w/o Eligibility trace | 58.3 | –21.8 | 60.2 | –22.4 |
| w/o Homeostasis | 74.2 | –5.9 | 68.4 | –14.2 |
Component-wise ablation study.
Δ indicates the relative performance drop compared to the full AGMP model.
5 Conclusion
In this work, we have addressed the critical challenges of memory efficiency and catastrophic forgetting in the training of deep Spiking Neural Networks by introducing Astrocyte-Gated Multi-Timescale Plasticity (AGMP). By theoretically extending the three-factor plasticity paradigm to incorporate a slow, biologically inspired astrocytic state, AGMP harmonizes fast temporal credit assignment with slow, stability-inducing regulation. Our results demonstrate that AGMP not only achieves scalable online learning with constant memory complexity but also provides robust continual learning capabilities without relying on privacy-invasive replay buffers.
This research bridges a significant gap between computational neuroscience and neuromorphic engineering, suggesting that non-neuronal cells, often overlooked in ANN design, play a pivotal role in solving the stability-plasticity dilemma. Looking forward, several promising directions emerge. While we utilized a global error broadcast, exploring more biologically plausible feedback mechanisms, such as local dendritic error prediction, could further enhance locality. Additionally, adapting AGMP for emerging Transformer-based SNN architectures or deploying it on FPGA-based neuromorphic accelerators could validate its real-time learning capabilities in large-scale, closed-loop scenarios. In summary, AGMP represents a significant step toward autonomous, energy-efficient intelligence capable of continuous lifelong adaptation.
Statements
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.
Author contributions
ZD: Conceptualization, Investigation, Methodology, Visualization, Writing – original draft. WH: Formal analysis, Methodology, Writing – review & editing.
Funding
The author(s) declared that financial support was received for this work and/or its publication. This work was supported in part by the National Science Foundation of Fujian Province under Grant 2023J05251.
Conflict of interest
WH was employed by Metatop (Fuzhou) Technology Co., Ltd.
The remaining author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declared that generative AI was not used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1
Amir A. Taba B. Berg D. Melano T. McKinstry J. Di Nolfo C. et al . (2017). “A low power, fully event-based gesture recognition system,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Honolulu, HI: IEEE). doi: 10.1109/CVPR.2017.781
2
Apolinario M. P. E. Roy K. (2024). S-tllr: Stdp-inspired temporal local learning rule for spiking neural networks. Trans. Mach. Learn. Res. arXiv:2306.15220.
3
Araque A. Parpura V. Sanzgiri R. P. Haydon P. G. (1999). Tripartite synapses: glia, the unacknowledged partner. Trends Neurosci. 22, 208–215. doi: 10.1016/S0166-2236(98)01349-6
4
Bellec G. Scherr F. Subramoney A. Hajek E. Salaj D. Legenstein R. et al . (2020). A solution to the learning dilemma for recurrent networks of spiking neurons. Nat. Commun. 11:3625. doi: 10.1038/s41467-020-17236-y
5
Benna M. K. Fusi S. (2016). Computational principles of synaptic memory consolidation. Nat. Neurosci. 19, 1697–1706. doi: 10.1038/nn.4401
6
Bernardinelli Y. Randall J. Janett E. Nikonenko I. König S. Jones E. V. et al . (2014). Activity-dependent structural plasticity of perisynaptic astrocytic domains promotes excitatory synapse stability. Curr. Biol. 24, 1679–1688. doi: 10.1016/j.cub.2014.06.025
7
Bohnstingl T. Woźniak S. Pantazi A. Eleftheriou E. (2023). Online spatio-temporal learning in deep neural networks. IEEE Trans. Neural Netw. Learn. Syst. 34, 8894–8908. doi: 10.1109/TNNLS.2022.3153985
8
Chen K. Li Z. (2024). Astrocyte mediated firing activities and synchronization in a heterogeneous neuronal network. Chaos Solitons Fractals188:115564. doi: 10.1016/j.chaos.2024.115564
9
Cramer B. Stradmann Y. Schemmel J. Zenke F. (2020). The heidelberg spiking data sets for the systematic evaluation of spiking neural networks. IEEE Trans. Neural Netw. Learn. Syst. 33, 2744–2757. doi: 10.1109/TNNLS.2020.3044364
10
Dan Y. Poo M.-M. (2004). Spike timing-dependent plasticity of neural circuits. Neuron44, 23–30. doi: 10.1016/j.neuron.2004.09.007
11
Diehl P. U. Neil D. Binas J. Cook M. Liu S.-C. Pfeiffer M. (2015). “Fast-classifying, high-accuracy spiking deep networks through weight and threshold balancing,” in 2015 International Joint Conference on Neural Networks (IJCNN) (Killarney: IEEE), 1–8. doi: 10.1109/IJCNN.2015.7280696
12
Frémaux N. Gerstner W. (2016). Neuromodulated spike-timing-dependent plasticity, and theory of three-factor learning rules. Front. Neural Circuits9:85. doi: 10.3389/fncir.2015.00085
13
Fusi S. Drew P. J. Abbott L. F. (2005). Cascade models of synaptically stored memories. Neuron45, 599–611. doi: 10.1016/j.neuron.2005.02.001
14
Gallego G. Delbrück T. Orchard G. Bartolozzi C. Taba B. Censi A. et al . (2022). Event-based vision: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 44, 154–180. doi: 10.1109/TPAMI.2020.3008413
15
Han B. Zhao F. Pan W. Zhao Z. Li X. Kong Q. et al . (2023a). Adaptive reorganization of neural pathways for continual learning with spiking neural networks. arXiv preprint arXiv:2309.09550. doi: 10.48550/arXiv.2309.09550
16
Han B. Zhao F. Zeng Y. Pan W. Shen G. (2023b). “Enhancing efficient continual learning with dynamic structure development of spiking neural networks,” in Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence (IJCAI) (Macao), 3787–3795. doi: 10.24963/ijcai.2023/334
17
Kirkpatrick J. Pascanu R. Rabinowitz N. Veness J. Desjardins G. Rusu A. A. et al . (2017). Overcoming catastrophic forgetting in neural networks. Proc. Natl. Acad. Sci. U. S. A. 114, 3521–3526. doi: 10.1073/pnas.1611835114
18
Li Y. Zeng X. Xue Z. Zeng P. Zhang Z. Wang Y. (2025). Incorporating the refractory period into spiking neural networks through spike-triggered threshold dynamics. arXiv preprint arXiv:2509.17769. doi: 10.1145/3746027.375503
19
Li Z. Hoiem D. (2018). Learning without forgetting. IEEE Trans. Pattern Anal. Mach. Intell. 40, 2935–2947. doi: 10.1109/TPAMI.2017.2773081
20
Lin E. Luo W. Jia W. Yang S. (2025). Online continual learning via spiking neural networks with sleep enhanced latent replay. arXiv preprint arXiv:2507.02901. doi: 10.48550/arXiv.2507.02901
21
Lopez-Paz D. Ranzato M. (2017). Gradient episodic memory for continual learning. Adv. Neural Inform. Process. Syst. 30, 6470–6479. doi: 10.48550/arXiv.1706.08840
22
Maass W. (1997). Networks of spiking neurons: the third generation of neural network models. Neural Netw. 10, 1659–1671. doi: 10.1016/S0893-6080(97)00011-7
23
Neftci E. O. Mostafa H. Zenke F. (2019). Surrogate gradient learning in spiking neural networks: bringing the power of gradient-based optimization to spiking neural networks. IEEE Sig. Process. Magaz. 36, 51–63. doi: 10.1109/MSP.2019.2931595
24
Orchard G. Jayawant A. Cohen G. K. Thakor N. (2015). Converting static image datasets to spiking neuromorphic datasets using saccades. Front. Neurosci. 9:437. doi: 10.3389/fnins.2015.00437
25
Parisi G. I. Kemker R. Part J. L. Kanan C. Wermter S. (2019). Continual lifelong learning with neural networks: a review. Neural Netw. 113, 54–71. doi: 10.1016/j.neunet.2019.01.012
26
Perea G. Navarrete M. Araque A. (2009). Tripartite synapses: astrocytes process and control synaptic information. Trends Neurosci. 32, 421–431. doi: 10.1016/j.tins.2009.05.001
27
Pes L. Yin B. Stuijk S. Corradi F. (2026). Traces propagation: memory-efficient and scalable forward-only learning in spiking neural networks. Neuromorphic Comput. Eng.6:014002. doi: 10.1088/2634-4386/ae2ef9
28
Porr B. Wörgötter F. (2007). Learning with “relevance”: using a third factor to stabilize hebbian learning. Neural Comput. 19, 2694–2719. doi: 10.1162/neco.2007.19.10.2694
29
Rastogi M. Lu S. Islam N. Sengupta A. (2021). On the self-repair role of astrocytes in stdp enabled unsupervised snns. Front. Neurosci. 14:603796. doi: 10.3389/fnins.2020.603796
30
Rathi N. Roy K. (2023). Diet-snn: a low-latency spiking neural network with direct input encoding and leakage and threshold optimization. IEEE Trans. Neural Netw. Learn. Syst. 34, 3174–3182. doi: 10.1109/TNNLS.2021.3111897
31
Rebuffi S.-A. Kolesnikov A. Sperl G. Lampert C. H. (2017). “icarl: incremental classifier and representation learning,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Honolulu, HI: IEEE), 2001–2010. doi: 10.1109/CVPR.2017.587
32
Reichenbach A. Derouiche A. Kirchhoff F. (2010). Morphology and dynamics of perisynaptic glia. Brain Res. Rev. 63, 11–25. doi: 10.1016/j.brainresrev.2010.02.003
33
Roy K. Jaiswal A. Panda P. (2019). Towards spike-based machine intelligence with neuromorphic computing. Nature575, 607–617. doi: 10.1038/s41586-019-1677-2
34
Rusu A. A. Rabinowitz N. C. Desjardins G. Soyer H. Kirkpatrick J. Kavukcuoglu K. et al . (2022). Progressive neural networks. arXiv preprint arXiv:1606.04671. doi: 10.48550/arXiv.1606.04671
35
Sengupta A. Ye Y. Wang R. Liu C. Roy K. (2019). Going deeper in spiking neural networks: Vgg and residual architectures. Front. Neurosci. 13:95. doi: 10.3389/fnins.2019.00095
36
Shan Y. Zhang M. Zhu R. Qiu X. Eshraghian J. K. Qu H. (2025). “Advancing spiking neural networks towards multiscale spatiotemporal interaction learning,” in AAAI Conference on Artificial Intelligence (Philadelphia, PA).
37
Shen G. Zhao D. Dong Y. Li Y. Li J. Sun K. et al . (2023). Astrocyte-enabled advancements in spiking neural networks for large language modeling. arXiv preprint arXiv:2312.07625. doi: 10.48550/arXiv.2312.07625
38
Shrestha S. B. Orchard G. (2018). “Slayer: spike layer error reassignment in time,” in Advances in Neural Information Processing Systems, vol. 31, eds. S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (Brookline, MA; Cambridge, MA: Curran Associates, Inc.), 1419–1428.
39
Song Y. Zhang A. (2025). From black box to physically interpretable: trustworthy computing for ai-driven decision-making and control. J. Autom. Intell. doi: 10.1016/j.jai.2025.09.003
40
Stogsdill J. A. Eroglu C. (2017). The interplay between neurons and glia in synapse development and plasticity. Curr. Opin. Neurobiol. 42, 1–8. doi: 10.1016/j.conb.2016.09.016
41
van de Ven G. M. Tolias A. S. (2019). Three scenarios for continual learning. arXiv preprint arXiv:1904.07734. doi: 10.48550/arXiv.1904.07734
42
Wade J. McDaid L. J. Harkin J. Crunelli V. Kelso S. (2012). Self-repair in a bidirectionally coupled astrocyte-neuron (an) system based on retrograde signaling. Front. Comput. Neurosci. 6:26578. doi: 10.3389/fncom.2012.00076
43
Wang C.-M. Dong X.-S. Jiang J. Ji Z. Liu X. Wu S. (2024). Brainscale: enabling scalable online learning in spiking neural networks. bioRxiv. doi: 10.1101/2024.09.24.614728
44
Warden P. (2018). Speech commands: a dataset for limited-vocabulary speech recognition. arXiv preprint arXiv:1804.03209. doi: 10.48550/arXiv.1804.03209
45
Wu K. Shunzhuo E. Yang N. Zhang A. Yan X. Mu C. et al . (2025). A novel approach to enhancing biomedical signal recognition via hybrid high-order information bottleneck driven spiking neural networks. Neural Netw. 183:106976. doi: 10.1016/j.neunet.2024.106976
46
Wu Y. Deng L. Li G. Zhu J. Shi L. (2018). Spatio-temporal backpropagation for training high-performance spiking neural networks. Front. Neurosci. 12:331. doi: 10.3389/fnins.2018.00331
47
Xiao M. Meng Q. Zhang Z. He D. Lin Z. (2022a). Online training through time for spiking neural networks. Adv. Neural Inform. Process. Syst. 35, 20717–20730.
48
Xiao M. Meng Q. Zhang Z. He D. Lin Z. (2022b). “Online training through time for spiking neural networks,” in Advances in Neural Information Processing Systems (NeurIPS) (New Orleans, LA).
49
Yin B. Corradi F. Bohté S. M. (2020). “Effective and efficient computation with multiple-timescale spiking recurrent neural networks,” in International Conference on Neuromorphic Systems (Oak Ridge, TN), 1–8. doi: 10.1145/3407197.3407225
50
Zenke F. Bohté S. M. Clopath C. Comşa I. M. Göltz J. Maass W. et al . (2021). Visualizing a joint future of neuroscience and neuromorphic engineering. Neuron109, 571–575. doi: 10.1016/j.neuron.2021.01.009
51
Zenke F. Ganguli S. (2018). Superspike: supervised learning in multilayer spiking neural networks. Neural Comput. 30, 1514–1541. doi: 10.1162/neco_a_01086
52
Zenke F. Poole B. Ganguli S. (2017). “Continual learning through synaptic intelligence,” in Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, eds. D. Precup and Y. W. Teh (Red Hook, NY: PMLR), 3987–3995.
53
Zhang A. Han Y. Niu Y. Gao Y. Chen Z. Zhao K. (2022). Self-evolutionary neuron model for fast-response spiking neural networks. IEEE Trans. Cogn. Dev. Syst. 14, 1766–1777. doi: 10.1109/TCDS.2021.3139444
54
Zhang A. Shi J. Wu J. Zhou Y. Yu W. (2024). Low latency and sparse computing spiking neural networks with self-driven adaptive threshold plasticity. IEEE Trans. Neural Netw. Learn. Syst. 35, 17177–17188. doi: 10.1109/TNNLS.2023.3300514
Summary
Keywords
astrocytes, continual learning, eligibility traces, neuromorphic computing, online learning, spiking neural networks, synaptic plasticity, three-factor learning
Citation
Dong Z and He W (2026) Astrocyte-gated multi-timescale plasticity for online continual learning in deep spiking neural networks. Front. Neurosci. 19:1768235. doi: 10.3389/fnins.2025.1768235
Received
16 December 2025
Revised
26 December 2025
Accepted
29 December 2025
Published
27 January 2026
Volume
19 - 2025
Edited by
Qichun Zhang, Buckinghamshire New University, United Kingdom
Reviewed by
Hui Gao, Shaanxi University of Science and Technology, China
Wangbin Ding, Fujian Medical University, China
Updates
Copyright
© 2026 Dong and He.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Wude He, wude.he@metatop.tech
Disclaimer
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.