# The neuronal response at extended timescales: a linearized spiking input–output relation

- Laboratory for Network Biology Research, Department of Electrical Engineering, Technion, Haifa, Israel

Many biological systems are modulated by unknown slow processes. This can severely hinder analysis – especially in excitable neurons, which are highly non-linear and stochastic systems. We show the analysis simplifies considerably if the input matches the sparse “spiky” nature of the output. In this case, a linearized spiking Input–Output (I/O) relation can be derived semi-analytically, relating input spike trains to output spikes based on known biophysical properties. Using this I/O relation we obtain closed-form expressions for all second order statistics (input – internal state – output correlations and spectra), construct optimal linear estimators for the neuronal response and internal state and perform parameter identification. These results are guaranteed to hold, for a general stochastic biophysical neuron model, with only a few assumptions (mainly, timescale separation). We numerically test the resulting expressions for various models, and show that they hold well, even in cases where our assumptions fail to hold. In a companion paper we demonstrate how this approach enables us to fit a biophysical neuron model so it reproduces experimentally observed temporal firing statistics on days-long experiments.

## 1. Introduction

Neurons are modeled biophysically using Conductance-Based Models (CBMs). In CBMs, the membrane time constant and the timescales of fast channel kinetics determine the timescale of Action Potential (AP) generation in the neuron. These are typically around 1–20 ms. However, there are various modulating processes that affect the response on slower timescales. Many types of ion channels exist, and some change with a timescale as slow as 10 s (Channelpedia), and possibly even minutes (Toib et al., 1998). Additional new sub-cellular kinetic processes are being discovered at an explosive rate (Bean, 2007; Sjöström et al., 2008; Debanne et al., 2011). This variety is particularly large for very slow processes (Marom, 2010).

For example, ion channels are known to be regulated over the course of long timescales (Levitan, 1994; Staub et al., 1997; Jugloff, 2000; Monjaraz et al., 2000), which could cause changes in ion channel numbers, conductances and kinetics. Also, the ionic concentrations in the cell depend on the activity of the ionic pumps, which can be affected by the metabolism of the network (Silver et al., 1997; Kasischke et al., 2004). Finally, the cellular neurites (De Paola et al., 2006; Nishiyama et al., 2007) and even the spike initiation region (Grubb and Burrone, 2010) can shift their location with time. All these changes can have a large effect on excitability.

Therefore, current CBMs can be considered as strictly accurate only below a certain timescale, since they do not incorporate most of these slow processes. A main reason for this “neglect” is that such slow processes are not well characterized. This is especially problematic since neurons are excitable, so their dynamics is far from equilibrium, highly non-linear and contain feedback. Due to the large number of processes which are unknown or lacking known parameters, it would be hard to simulate or analyze such models. Therefore, it may be hard to quantitatively predict how adding and tuning slow processes in the model would affect the dynamics at longer timescales.

In order to allow CBMs with many slow process to be fitted and analyzed, it is desirable to have general expressions that describe their Input–Output (I/O) relation explicitly, based on biophysical parameters. In a previous paper (Soudry and Meir, 2012b), we found that this becomes possible if we use (experimentally relevant Elul and Adey, 1966; Kaplan et al., 1996; De Col et al., 2008; Gal et al., 2010; Goldwyn et al., 2012) sparse spike inputs, similar to the typical output of the neuron (Figures 1A,B). In this case, we derived semi-analytically^{1} a discrete piecewise linear map describing the neuronal dynamics between stimulation spikes, for a general *deterministic* neuron model with a few assumptions (mainly, a timescale separation assumption). Based on this reduced map, we were able to derive expressions for the “mean” behavior of the neuron (e.g., firing modes, firing rate and mean latency).

**Figure 1. Schematic summary. (A)** Aim: find the I/O relation between inter-stimulus intervals (*T*_{m}) and Action Potential (AP) occurrences (*Y*_{m}) – for a *general* biophysical neuron model (Equations 1–3). **(B)** An AP “occurred” if the voltage *V* crossed a threshold *V*_{th} following the (sparse) stimulus, with *T*_{m} ≫ τ_{AP}. **(C)** Result: Biophysical neuron model reduced to a simple linear system with feedback (Equations 11, 12), and biophysically meaningful parameters (**F,d,a** and **w**).

In this paper, we find that stronger and more general analytical results can be obtained if we take into account the stochasticity of the neuron – arising from ion channel noise^{2} (Neher and Sakmann, 1976; Hille, 2001). Due to the presence of this noise, the discrete map describing the neuronal dynamics is “smoothed out,” and can be linearized. This linearized map constitutes a concise description for the neuronal I/O (Equations 11, 12) based on biophysically meaningful parameters. This I/O is well described by an “engineering-style” block diagram with feedback (Figure 1C), where the input is the process of stimulation intervals and the output is the AP response (Figure 1A). Note that the response is affected both by internal noise and by the input. Beyond conceptual lucidity, such a linear I/O allows the utilization of well known statistical tools to derive all second order statistics, to construct linear optimal estimators and to perform parameter identification. These results hold numerically (Figure 3), even sometimes when our assumptions break down (Figure 4).

In our previous paper, Soudry and Meir (2012b), we used our results to model recent experiments (Gal et al., 2010) where synaptically isolated individual neurons, from rat cortical culture, were stimulated with extra-cellular sparse current pulses for an unprecedented duration of days. Our results enabled us to explain the “mean” response of these neurons. However, the second order-statistics in the experiment seem particularly puzzling. The neurons exhibited 1/*f*^{α} statistics (Keshner, 1982), responding in a complex and irregular manner from seconds to days. In a companion paper (Soudry and Meir, 2014), we demonstrate the utility of our new results. These results allow us to reproduce and analyze the origins of this 1/*f*^{α} on very long timescales.

## 2. Results

This section described our main results in outline. The details of each sub-section here appear in the corresponding sub-section in section 4. For readers who do not wish to go through the detailed derivations, the present section is self-contained. Readers who do wish to follow the mathematical derivations, should first read section 4, where, for convenience, each subsection (except for the last one) can be read independently. In our notation 〈·〉 is an ensemble average, ${i}{\triangleq}\sqrt{{-}{1}}$, a non-capital boldfaced letter **x** ≜ (*x*_{1}, …, *x*_{n})^{⊤} is a column vector (where (·)^{⊤} denotes transpose), and a boldfaced capital letter **X** is a matrix (with components *X*_{mn}).

### 2.1. Full Model

The voltage dynamics of an isopotential neuron are determined by ion channels, protein pores which change their conformations stochastically with voltage-dependent rates (Hille, 2001). At the population level, such dynamics are generically described by Fox and Lu (1994), Goldwyn et al. (2011), and Soudry and Meir (2012b) a CBM

with **voltage** *V*, stimulation current *I*(*t*), **rapid** variables **r** (e.g., *m, n, h* in the Hodgkin–Huxley (HH) model Hodgkin and Huxley, 1952), **slow** “excitability” variables **s** (e.g., slow sodium inactivation Chandler and Meves, 1970), rate matrices **A**_{r/s}, white noise processes ξ_{r/s} (with zero mean and unit variance), and matrices **B**_{r/s} which can be written explicitly using the rates and ion channel numbers (Orio and Soudry, 2012) (**D** = **BB**^{⊤} is the diffusion matrix Orio and Soudry, 2012). For simplicity, we assumed that **r** and **s** are not coupled directly, but this is non-essential (Contou-Carrere, 2011; Wainrib et al., 2011). The parameter space can be constrained (Soudry and Meir, 2012b), since we consider here only *excitable*, non-oscillatory neurons which do not fire spontaneously^{3} and which have a single resting state – as is common for cortical cells, e.g., Gal et al. (2010).

Since the components of **r** and **s** usually represent fractions, in some cases it is more convenient to use the normalization constraint (i.e., that fractions sum to one), and reduce the dimensions of **r, s**, and ξ_{r/s}. After this reduction, the form of Equations (1–3) changes to

where all the variables and parameters have been redefined (with their size decreased). Note that we have slightly abused notation by using the same symbols in Equations (4–6) and in Equations (1–3). The specific set of equations used will always be stated. We call Equations (4–6) the “compressed form” of the CBM.

Such biophysical neuronal models (either Equations 1–3 or 4–6) are generally complex and non-linear, containing many variables and unknown parameters (sometimes ranging in the hundreds Koch and Segev, 1989; Roth and Häusser, 2001), not all of which can be identified (Huys et al., 2006). Therefore, such models are notoriously difficult to tune, highly susceptible to over-fitting and computationally expensive (Migliore et al., 2006; Gerstner and Naud, 2009; Druckmann et al., 2011). Also, the high degree of non-linearity usually prevents exact mathematical analysis of such models at their full level of complexity (Ermentrout and Terman, 2010). However, much of the complexity in such models can be overcome under well defined and experimentally relevant settings (Elul and Adey, 1966; Kaplan et al., 1996; De Col et al., 2008; Gal et al., 2010; Goldwyn et al., 2012), if we use sparse inputs, similar in nature to the spikes commonly produced by the neuron.

### 2.2. Model Reduction

We consider a stimulation setting motivated by the experiments described in Gal et al. (2010) and further elaborated on in section 3. Specifically, suppose *I*(*t*) consists of a train of pulses arriving at times {*t*_{m}} (Figure 1A, *top*), so *T*_{m} = *t*_{m + 1} − *t*_{m} ≫ τ_{AP} with τ_{AP} being the timescale of an AP (Figure 1B). Our aim is to describe the AP occurrences *Y*_{m}, where *Y*_{m} = 1 if an AP occurred immediately after the *m*-th stimulation, and 0 otherwise (Figure 1A, *bottom*). Recall again that we assume the neuron does not generate APs unless stimulated (as observed in Gal et al., 2010).

In this section we “average out” Equations (1–3) using a semi-analytical method similar to that in Soudry and Meir (2012b). To do so, we need to integrate Equations (1–3) between *t*_{m} and *t*_{m + 1}. Since *T*_{m} ≫ τ_{AP}, the rapid AP generation dynamics of (*V*, **r**) relax to a steady state before *t*_{m + 1}. Therefore, the neuron AP “remembers” any history before *t*_{m} only through **s**_{m} = **s**(*t*_{m}). Given **s**_{m}, the response of the fast variables (*V*, **r**) to the *m*-th stimulation spike will determine the probability to generate an AP. This probability,

collapses all the relevant information from Equations (1, 2), and can be found numerically from the pulse response of Equations (1, 2) with **s** held fixed (section 4.2.4).

In order to integrate the remaining Equation (3), we define *X*_{+}, *X*_{−} and *X*_{0} to be the averages of a quantity *X*_{s} during an AP response, a failed AP response and rest, respectively ^{4}. Also, we denote

as the steady state mean value of *X*_{s}. For analytical simplicity we assume^{5} *T*_{m} ≪ τ** _{s}**. We obtain, to first order

where **n**_{m} is a white noise process with zero mean and variance *T*_{m}**D** (*Y*_{m}, *T*_{m}). For the compressed form (Equations 4–6) we have instead

Note that such a simplified discrete time map, which describes the excitability dynamics of the neuron, has far fewer parameters than the full model, since it is written explicitly only using the averaged microscopic rates of **s** (through **A** and **D**), population sizes (through **D**), the probability to generate an AP given **s**, *p*_{AP} (**s**), and the relevant timescales. This effective model exposes the large degeneracy in the parameters of the full model and leads to significantly reduced simulation times and mathematical tractability. Notably, the dynamics of the state **s**_{m} (Equation 8) depends on the input *T _{m} and* the output

*Y*

_{m}– and this feedback affects all of our following results.

### 2.3. Linearization

In this section we exploit the intrinsic ion channel noise to linearize the neuronal dynamics, rendering it more tractable than the (less realistic) noiseless case (Soudry and Meir, 2012b). Suppose that the inter-stimulus intervals {*T*_{m}} have stationary statistics with mean *T*_{*} so that τ_{AP} ≪ *T*_{m} ≪ τ** _{s}** with high probability. Since

**s**is slow and AP generation is rather noisy in this regime (Soudry and Meir, 2012b) (so

*p*

_{AP}(

**s**

_{m}) is slowly varying), we assume that a stable excitability fixed point

**s**

_{*}exists (Figure 2). Therefore, the perturbations ${\widehat{{s}}}_{{m}}{=}{{s}}_{{m}}{-}{{s}}_{{\ast}}$ are small and we can linearize

**Figure 2. Schematic explanation of linearization**. In a deterministic neuron, an AP will be generated in response to stimulation if and only if the neuronal excitability (here, **s**) is above a certain threshold **(A)**. This generates discontinuous dynamics in the neuronal excitability **(B)**, see Equations 7, 8). In a stochastic neuron, the response probability is a smooth function of **s (C)**. In turn, this “smooths” the dynamics **(D)**. Note that if the noise is sufficiently high (as is true in many cases, for biophysically realistic levels of noise), then this generates a stable fixed point **s*** – which gives the mean response probability *p*_{*}, and around which the dynamics can be linearized (yellow region).

Denoting *X*_{*} = *X*(*p*_{*}, *T*_{*}), the mean AP firing rate can be found self consistently from the location of the fixed point **s**_{*},

where **s**_{*} depends on *p*_{*} through **A**_{*}**s**_{*} = 0, or **s**_{*} = **A**^{−1}_{*}**b**_{*} in the compressed form.

The perturbations ${\widehat{{s}}}_{{m}}{=}{{s}}_{{m}}{-}{{s}}_{{\ast}}$ around the fixed point **s**_{*} are described by the linear system

where ${\widehat{{T}}}_{{m}}{=}{{T}}_{{m}}{-}{{T}}_{{\ast}}$, ${\widehat{{Y}}}_{{m}}{=}{{Y}}_{{m}}{-}{\langle}{{Y}}_{{m}}{\rangle}$, **F** ≜ **I** + *T*_{*}**A**_{*}, 〈**n**_{m}**n**^{⊤}_{m}〉 = *T*_{*}**D**_{*}, *e*_{m} is a (*non*-Gaussian) white noise process, 〈*e*_{m}〉 = 〈*e*_{m}**n**_{m}〉 = 0, σ^{2}_{e} ≜ 〈*e*^{2}_{m}〉 = *p*_{*} (1 − *p*_{*}), **d** ≜ **A**_{0}**s**_{*} and **a** ≜ τ_{AP} (**A**_{+} − **A**_{−}) **s**_{*}. If we use the compressed form instead, then these results remain valid, except we need to re-define **d** ≜ **A**_{0}**s**_{*} − **b**_{0} and **a** ≜ τ_{AP}[(**A**_{+} − **A**_{−}) **s**_{*} − (**b**_{+} − **b**_{−})].

The *linear* I/O for the fluctuations in Equations (11, 12), which contains feedback from the “output” ${\widehat{{Y}}}_{{m}}$ to the state variable ${\widehat{{s}}}_{{m}}$ (Figure 1C), can be very helpful mathematically and its parameters are directly related to biophysical quantities.

### 2.4. Linear Systems Analysis

Using standard tools, this formulation makes it now possible to construct optimal linear estimators for *Y*_{m} and **s**_{m} (Anderson and Moore, 1979), perform parameter identification (Lejung, 1999), and find all second order statistics in the system (Papoulis and Pillai, 1965; Gardiner, 2004), such as correlations or Power Spectral Densities (PSD). For example, for *f* ≪ *T*^{−1}_{*}, the PSD of the output is

where

Similarly, the PSD of the state variables is

and the input–output cross-PSD is

Again, note the large degeneracy here – many different sets of parameters will generate the same PSD. Using similar methods, the PSDs of various response features, such as the AP latency or amplitude, can also be derived (Equation 124).

Finally, we note Equations (11) and (12) can be re-arranged as a *direct* I/O relation. First, we define the filters (transfer functions)

where **K** = **a** + **FPw**σ^{−2}_{v} and σ^{2}_{v} = **w**^{⊤} **Pw** + σ^{2}_{e}, with **P** being the solution of

Using these filters, we obtain, in the frequency domain,

where $\widehat{{Y}}{\left(}{f}{\right)}{,}\widehat{{T}}{\left(}{f}{\right)}$ and *z*(*f*) are the Fourier transforms of *Y*_{m}, ${\widehat{{T}}}_{{m}}$ and *z*_{m}, respectively, with *z*_{m} being a white noise process with zero mean and unit variance. Notably, these transfer functions can be identified from the spiking input–output of the neuron ${\left\{}{\widehat{{T}}}_{{m}}{,}{\widehat{{Y}}}_{{m}}{\right\}}$, without access to the underlying dynamics or biophysical parameters. Specifically, Equation (20) has the form of an ARMAx(*M, M, M*) model^{6} (Lejung, 1999) (recall *M* is the dimension of **s**), which can be estimated using standard tools (e.g., the system identification toolbox in Matlab).

### 2.5. Numerical Tests

As we argued so far, a main asset of the present approach is its applicability to a broad range of models of various degrees of complexity and realism. Recall that our three assumptions are

- τ
_{AP}≪*T*_{m}(temporally sparse input). *T*_{m}≪ τ(timescale separation)._{s}- A stable excitability fixed point
**s**_{*}exists, (“noisy” neuron).

In this section we will demonstrate that our analytical approximations agree very well with the numerical solution of Equations (1–3), even in some cases where the assumptions 2 and 3 do not hold. Therefore, these assumptions are sufficient, but not necessary.

#### 2.5.1. The HHS model

First, in Figure 3 we tested our results on the HH model with Slow sodium inactivation. This “HHS” model (Soudry and Meir, 2012b, and see section 4.5.1 for parameter values) augments the classic HH model (Hodgkin and Huxley, 1952) with an additional slow inactivation process of the sodium conductance (Chandler and Meves, 1970; Fleidervish et al., 1996). The HHS model includes the uncoupled stochastic Hodgkin–Huxley (HH) model equations (Fox and Lu, 1994), and is written in the compressed formulation (Equations 4–6)

for *r* = *m, n* and *h*, with the additional kinetic equation for slow sodium inactivation

where *V* is the membrane voltage, *I*(*t*) is the input current, *m, n* and *h* are ion channel “gating variables,” α_{r}(*V*), β_{r}(*V*), δ(*V*), and γ(*V*) are the voltage dependent kinetic rates of these gating variables, ϕ is an auxiliary dimensionless number, *C* is the membrane's capacitance, *E*_{K}, *E*_{Na} and *E*_{L} are ionic reversal potentials, *g*_{K}, *g*_{Na} and *g*_{L} are ionic conductances and *N* is the number of ion channels. Note that in this model τ_{s} is between 20 s (at rest) and 40 s (during an AP).

**Figure 3. Comparing the mathematical results with the numerical simulation of the full model (Equations 1–3) for the stochastic HHS model (section 4.5.1)**. **(A)** Firing probability *p*_{*}(*T*^{−1}_{*}) (Equation 10) for different currents (*I*_{stim} = 7.5, 7.7, 7.9, 8.1, 8.3 μA from bottom to top). **(B)** The PSDs *S*_{Y}(*f*) and *S*_{s}(*f*). “Sim” is a simulation of the full model, “Map” is a (10^{4} faster) simulation of Equation (8) together with *p*_{AP} (**s**_{m}), while “Approx” refers to the analytical expressions (Equations 13–15). “Ident” is the PSD *S*_{Y}(*f*) of the linear system identified from the spiking data. Note the high/low-pass filter shapes of *S*_{Y}(*f*) and *S*_{s}(*f*), respectively. **(C)** Optimal linear estimation of $\widehat{{s}}$. **(D)** Amplitude and phase of the cross-spectrum *S*_{YT}(*f*) for Poisson stimulation (Equations 16). Note that the frequency range was cut due to spectral estimation noise (see Figure 8). Parameters: *I*_{0} = 7.9 μA and *T*_{*} = 50 ms in **(B–D)**, and also stimulation is periodical in **(A–C)**. Note the low-pass filter shapes of *S*_{YT}(*f*).

In Figure 3A we show that through Equation (10) we can accurately calculate *p*_{*}, the mean probability to generate an AP (so *p*_{*}*T*^{−1}_{*} is the firing rate of the neuron). In Figure 3B we demonstrate both the analytical expression (Equations 13, 15), or a simulation of the reduced model (Equation 8), will give the PSDs *S*_{Y} (*f*) or *S*_{s} (*f*) of the full model (Equations 1–3). In Figure 3D we do the same for the analytical expression (Equation 16) of the Cross-PSD *S*_{YT} (*f*). In Figure 3C we show that we can construct a linear optimal filter for the internal state ${\widehat{{s}}}_{{m}}$, given ${\left\{}{{\left\{}{{T}}_{{k}}{\right\}}}_{{k}{\text{\hspace{0.05em}}}{=}{\text{\hspace{0.05em}}}{0}}^{{m}{\text{\hspace{0.05em}}}{-}{\text{\hspace{0.05em}}}{1}}{,}{{\left\{}{{Y}}_{{k}}{\right\}}}_{{k}{\text{\hspace{0.05em}}}{=}{\text{\hspace{0.05em}}}{0}}^{{m}{\text{\hspace{0.05em}}}{-}{\text{\hspace{0.05em}}}{1}}{\right\}}$ quite well, with low mean square error (section 4.4.4). Finally, back in Figure 3B, *top*, we infer the linear model parameters from the spike output using system identification tools [here, with ARMAx(1, 1, 1)], and present the PSD of the identified model (“Ident”). Since *S*_{Y} (*f*) = |*H*_{int} (*f*)|^{2} (see Equation 111) for periodical input (in which ${\widehat{{T}}}_{{m}}{=}{0}$) this allows us to confirm that the linear model was identified. As can be seen, the identified filter matches well with that of the linear system.

#### 2.5.2. Testing the limit of our assumptions

Next, we demonstrate that our analytical expressions hold also for various other models. Specifically, in the following scenarios: (1) when the kinetics of the neuron are extended to arbitrarily slow timescales, (2) when the assumptions 2 and 3 break down, (3) when the rapid and slow kinetics are coupled, (4) when “physiological” synaptic inputs are used. These results are presented in Figures 4, 5, with specific model parameters given in section 4.5.

**Figure 4. Comparing mathematical results with full model simulation when the assumptions fail to hold**. In the HHSIP model (HHS with potassium inactivation) we plot **(A)** *p*_{*}(*T*^{−1}_{*}) for different currents (*I*_{0} = 7.5, 7.7, 7.9, 8.1, 8.3μA from bottom to top). **(B)** *S*_{Y}(*f*) for two values of *T*_{*}. As before, “Sim” is a simulation of the full model, “Approx” is the analytical approximation, and “Ident” is the PSD *S*_{Y}(*f*) of the linear system identified from the spiking data. Upper figure shows the case when *T*_{*} ≈ 0.5τ** _{s}** so the timescale separation assumption breaks down. In the lower figure the parameters are close to a Hopf bifurcation where a limit cycle is formed so the fixed point assumption breaks down, so the estimation of the limit cycle frequency component is less accurate.

**(C)**The estimation of ${\widehat{{s}}}_{{1}}$ for

*T*

^{−1}

_{*}≈ 30 Hz is even better than in the HHS case. Similarly to

**(A–C)**we plot the results of the HHMSIP model (HHSIP with many additional slow sodium inactivation kinetics) in

**(D–F)**, which has considerably more noise in the slow kinetics, and so even larger fluctuations (which further invalidates the fixed point assumption). See section 4.5 for various model details.

**Figure 5. Comparing mathematical results (green) with full model simulation (blue) for various models. (A)** Coupled HHS (HHS coupled slow and rapid kinetics) **(B)** HHMS (HHS with many additional slow sodium inactivation kinetics) **(C)** HHSTM (HHS with a synapse) **(D)** Multiplicative HHMS (variant of HHMS). As before, “Sim” is a simulation of the full model, “Approx” is the analytical approximation, and “Ident” is the PSD *S*_{Y}(*f*) of the linear system identified from the spiking data. See section 4.5 for various model details.

First, we tested whether or not the model can be extended to arbitrarily slow timescales. We added to the HHS model four types of slow sodium inactivation processes with increasingly slower kinetics and smaller channel numbers. In the first case, those processes were added additively (as different currents), so *s* was replaced with ∑_{i} *s*_{i} in the voltage equation (Equation 21). This model was denoted “HHMS” (HH with Many Sodium slow inactivation processes, section 4.5.4). In the second case, those processes were added in a multiplicative manner (as different processes affecting the same channel, in the uncoupled approximation), so *s* was replaced with ∏_{i} *s*_{i} in the voltage equation (Equation 21). We denote this model as “Multiplicative HHMS” (section 4.5.5). In both cases, our analytical approximations seemed to hold quite well. For example, the approximated *S*_{Y} (*f*) (Equation 13) corresponded rather well with the numerical simulation of the full model (Figures 5B,D, respectively).

Next, to test the limits of our assumptions we extended the HHS model to the HHSIP model (from Soudry and Meir, 2012b, see section 4.5.6) and added a potassium inactivation current which had faster kinetics (so τ** _{s}** ≈ 5 Hz). So if

*T*

^{−1}

_{*}= 10 Hz, we get

*T*

_{*}≈ 0.5τ

**, so the timescale separation assumption 2 is not strictly valid here. Also, for certain parameter values we get a limit cycle in the dynamics of ${\widehat{{s}}}_{{m}}$, so the fixed point assumption 3 fails. However, it seems that our approximations still follow the numerical simulation of the full model: for**

_{s}*p*

_{*}at various stimulation frequencies

*T*

^{−1}

_{*}and currents

*I*

_{0}(Figure 4A), for

*S*

_{Y}(

*f*) at

*T*

^{−1}

_{*}= 10 Hz when assumption 2 breaks down (Figure 4B,

*top*), for

*S*

_{Y}(

*f*) at

*T*

^{−1}

_{*}= 30 Hz when assumption 3 breaks down (near a Hopf bifurcation) and a limit cycle begins to form (see Figure 4B,

*bottom*), and for state estimation of ${\widehat{{s}}}_{{1}}$ using a linear optimal filter, again at

*T*

^{−1}

_{*}= 10 Hz (Figure 4C).

The only discrepancy seemed to appear in the limit cycle case, where the frequency of the limit cycle “sharpens” the peak in *S*_{Y} (*f*) (Figure 4B, *bottom*). This may suggest that, in this case, the perturbations of the system near the limit cycle could be linearized, and that the eigenvalues of that linearized system might be related to the eigenvalues of the linearized system around the (now unstable) fixed point **s**_{*}. More generally, the results so far indicate that even if our assumptions are inaccurate, it is possible that the resulting error will not accumulate and remain small – in comparison with the intrinsic noise in the model.

Next, to challenge the approximation even more, we added to the HHSIP model four types of sodium currents with increasingly slower kinetics and fewer channels, similarly to the HHMS model (so this is the “HHMSIP” model, section 4.5.7). This significantly increased the variance of the dynamic noise **n**_{m}, rendering the dynamics more “noisy.” These random fluctuations in **s**_{m} (Figure 4E) are of similar magnitude to the width of the threshold (non-saturated) region in *p*_{AP} (**s**_{m}) (see Figure 6). This renders the fixed point assumption 3 inaccurate, since now the linear approximation ${{p}}_{{\text{AP}}}{\left(}{{s}}_{{m}}{\right)}{\approx}{{p}}_{{\ast}}{+}{{w}}^{{\top}}{\widehat{{s}}}_{{m}}$ breaks down most of the time. However, even in this case, the approximations seem to hold quite well with simulations of the full neuronal model (Figures 4D–F).

**Figure 6. Fitting of p_{AP}(s) = Φ((s − a)/b) in the HHS model. (A)** Fitting of

*p*

_{AP}(

*s*) for various values of

*I*

_{0}.

**(B)**Fitting shows that

*a*is linearly decreasing in

*I*

_{0}.

**(C)**Fitting of

*p*

_{AP}(

*s*) for various values of

*N*.

**(D)**Fitting shows that ${b}{\propto}{1}{/}\sqrt{{N}}$.

In Figure 5A we used a coupled version of the HHS model (“coupled HHS” model, section 4.5.2), in which the equations for **r** and **s** in the full model are tangled together, and not separated as we assumed in Equations (2, 3). Even in this case, our approximations seemed to hold well.

Finally, in Figure 5C, we extend the HHS model so that the stimulations are not given directly, but through a synapse. We used the biophysical Tsodyks–Markram model (Tsodyks and Markram, 1997) of a synapse with short-term depression, with added stochasticity (“HHSTM” model, section 4.5.3). This also seemed to work well.

In all simulation we also added the PSD of the linear model identified from the spike output (“Ident.”), to show that it can be estimated reasonably well. Note that the performance at the lowest frequencies seems to be significantly worse when they contain relatively high power. This is not surprising since it is typically harder to estimate model parameters, when the data has such (1/*f*^{α}) PSD shape – which indicates long-term correlations (Beran, 1992).

## 3. Discussion

In this work we found that under a temporally sparse (“spike-like”) stimulation regime (Figures 1A,B) we can perform accurate semi-analytical linearization of the spiking input–output relation of a CBM (Figure 1C), while retaining biophysical interpretability of the parameters (e.g., Figure 7). This linearization considerably reduces model complexity and parameter degeneracy, and enables the use of standard analysis and estimation tools. Importantly, this method is rather general, since it can be applied to any stochastic CBM, with only a few assumptions.

**Figure 7. The averaged kinetic rates**. **Left**: The averaged rates demonstrated for three common kinetic rates γ(*V*) with sigmoidal shapes. **Right**: The voltage threshold of the sigmoid determines whether the process is sensitive to APs (the output), stimulation pulse (the input), or neither. Note that a similar classification of biophysical processes affecting excitability was previously suggested in Wallach (2012, Figure 3.1).

### 3.1. Connection to Previous Work

To the best of our knowledge, such results are novel, as no previous work examined analytically the response of general stochastic CBMs to temporally sparse input for extended durations. However, the connection between sparse inputs and slow timescales has been previously made. It was previously suggested (Linaro et al., 2011) that sparse inputs could be used to identify neuronal parameters in a network of integrate and fire neurons with spike frequency adaptation. Interestingly, using different methods we reach a qualitatively similar conclusion here, though not in a network setting, and for a different class of neuron models.

Additionally, in Soudry and Meir (2012b) we modeled neurons under *periodical* stimulation using *deterministic* CBMs with slow kinetics, which are completely *uncoupled* from each other, and slower than the stimulation rate. Using a reduction scheme similar in nature to that described here, we were able to describe the deterministic CBM's excitability and response using a discrete-time map – which “samples” the neuronal state at each stimulation. Analyzing this map, we obtained analytical results describing the neuronal activation modes, spike latency dynamics, mean firing rate and short-time firing patterns. Stochastic CBMs were then examined numerically, and were shown to lead to qualitatively different responses, which are more similar to the experimentally observed responses.

The current work, therefore, generalizes this previous work. Here, we analyze the general case of *stochastic* CBMs, under *general* sparse stimulation patterns and with *coupled* slow kinetic dynamics. Therefore, the framework in the previous work (Soudry and Meir, 2012b) could be considered as a special case of this work, in which there is an infinite number of ion channels (*N* → ∞, so **B**_{r/s} = **D**_{r/s} = 0), *T*_{m} = *T*_{*} (so ${\widehat{{T}}}_{{m}}{=}{0}$) and **A**_{s} (*V*) (the rate matrix) is a diagonal matrix. In the current work we similarly show that, in the generalized framework, the CBM's excitability and responses can be succinctly described using a discrete-time map. It is then straightforward to derive results paralleling those in Soudry and Meir (2012b) in this more general setting, such as the mean firing rate (Equation 10).

### 3.2. Theoretical Novelty

However, the main novelty lies in our additional results, that could not be derived in Soudry and Meir (2012b). Specifically, due to the presence of noise, we were able to linearize the map's dynamics, and derive an explicit input–output relation. Such a linearization became possible because we made the (unusual) choice that the “input” to the CBM consists of the time-intervals between stimulation pulses, while the “output” is a binary series indicating whether or not an AP happened immediately after a stimulation pulse. The linearized input–output relation can be expressed either in biophysically interpretable “state space” (Equations 11, 12 and Figure 1C), or as a sum of the filtered input and filtered noise (Equation 20). Note that the overall I/O includes the mean output (Equation 10) which is non-linear. However, the linear part of the response, allows the derivation of the power spectral densities (Equation 13), the construction of linear optimal estimators (e.g., Figure 3C) and blind identification of the (linearized) system parameters (“Ident.” in Figures 3–5).

Our results rely on three main assumptions. The temporal sparseness of the input τ_{AP} ≪ *T*_{m} insures that the slow variables **s**_{m} effectively represent the “neuronal state” alone (as *V* and **r** always relax to a steady state before the next stimulation is given). The additional assumption *T*_{m} ≪ τ** _{s}** allowed us to integrate the model dynamics and derive the reduced map (Equation 8) for the dynamics of

**s**

_{m}, which is linear in

*T*

_{m}. The last assumption is that the dynamics of

**s**

_{m}can be linearized around a stable fixed point

**s**

_{*}. This fixed point is generated due to the noisiness of the rapid variables (Figure 2), and the assumption

*T*

_{m}≪ τ

**ensures that the stochastic fluctuations around**

_{s}**s**

_{*}are small. We performed extensive numerical simulations (section 2.5) that indicate that our analytical results are accurate – sometimes even if our assumptions break down.

However, clearly there are cases, beyond our assumptions, in which are results cannot hold. For example, if ${\widehat{{T}}}_{{m}}$ has very large fluctuations, then the response of the neuron cannot be completely linear, since ${0}{<}{\widehat{{Y}}}_{{m}}{<}{1}$. Such cases may require an extension of the formalism described here. There are many possible extensions which we did not pursue here. For example, one can extend the modeling framework (e.g., multi-compartment neurons), stimulation regime (e.g., heterogeneous pulse amplitudes), or the type of neurons modeled (e.g., bursting and spontaneously firing neurons). However, it seems that an important assumption, that cannot be easily removed, is that the input is temporally sparse (τ_{AP} ≪ *T*_{m}).

### 3.3. Practical Significance

Is such a sparse temporally stimulation regime “physiologically relevant” for the soma of a neuron? Currently, such question cannot be answered directly, since it is impossible to accurately measure all the current arriving to the soma from the dendrites under completely physiological conditions. However, there is some indirect evidence. Recent studies have shown that the distribution of synaptic efficacies in the cortex is log-normal (Song et al., 2005) – so a few synapses are very strong, while most are very weak. This indicates that the neuronal firing patterns might in fact be dominated by a small number of very strong synapses while the sum of the weak synapses sets the voltage baseline (Ikegaya et al., 2012). Such a possibility is supported by the fact that individual APs can trigger the complex network events in humans (Molnár et al., 2008; Komlósi et al., 2012). Also, in rats, individual cortical cells can elicit whisker movements in Brecht et al. (2004) and even modify the global brain state (Li et al., 2009). Taken together, these observations suggests that the above-threshold stimulation reaching the soma may be temporally sparse in some cases.

There are other obvious cases were our results are immediately applicable. First, in an axonal compartment, the relevant input current is indeed sparse – an AP spike train arriving from a previous compartment. Second, a direct pulse-like stimulation is used in cochlear implants (Goldwyn et al., 2012, and references therein). Lastly, such stimulation is used as an experimental probe (De Col et al., 2008; Gal et al., 2010; Gal and Marom, 2013). Specifically, since we now have a precise expression for the power spectral density of the response, we are now ready to use these analytical results in Soudry and Meir (2014) to reproduce the experimentally observed 1/*f*^{α} behavior of the neuron and explore its implications on its input–output relation.

## 4. Methods

In this section we provide the details of the results presented in the paper. Sections 4.1–4.5 here respectively correspond to Sections 2.1–2.5. The first four (theoretical) sections can be read independently of each other (except when we discuss the repeating “HHS model” example). The last section give the details of the numerical simulations.

### 4.1. Full Model (Biophysical Neuron Models)

As we explained in section **2.1**, a general model for a biophysical isopotential neuron is given by the following equations

with **voltage** *V*, stimulation current *I*(*t*), **rapid** variables **r** (e.g., *m, n, h* in the Hodgkin–Huxley (HH) model Hodgkin and Huxley, 1952), **slow** variables **s** (e.g., slow sodium inactivation Chandler and Meves, 1970), rate matrices **A**_{r/s}, white noise processes ξ_{r/s} (with zero mean and unit variance), and matrices **B**_{r/s} which can be written explicitly using the rates and ion channel numbers (Orio and Soudry, 2012) (**D** = **BB**^{⊤} is the diffusion matrix Gardiner, 2004; Orio and Soudry, 2012). In this section we give the specific forms of **A**_{r/s} and **B**_{r/s}, and their origin based on neuronal biophysics.

#### Microscopic origins

Such a model is commonly called a stochastic Conductance Based Model (CBM). In a non-stochastic CBM, the dynamics of the membrane voltage *V* (Equation 36) are deterministically determined by some general function of *V*, the stimulation current *I*(*t*), and some internal state variables **r** and **s**. In contrast, the dynamical equations for **r** and **s** here adhere to a specific Stochastic Differential Equation (SDE) form, since these variables describe the “population state” of all the ion channels in the neuron. We now explain the biophysical interpretation of those equations.

At the microscopic level, each ion channel has several states, and it switches between those states with voltage dependent rates (Hille, 2001). This is usually modeled using a Markov model framework (Colquhoun and Hawkes, 1981). Formally, suppose we index by *c* the different types of channels, *c* = 1, …, *C*. For each channel type *c* there exist *N*^{(c)} channels, where each channel of type *c* possesses *K*^{(c)} internal states. In the Markov framework, for each ion channel that resides in state *i*, the probability that the channel will be in state *j* after an infinitesimal time *dt* is given by

where **A**^{(c)} (*V*) is called the “rate matrix” for that channel type.

To facilitate mathematical analysis and efficient numerical simulation, we preferred to model stochastic CBMs using a compressed, SDE form. This method was initially suggested by Fox and Lu (1994), but their method suffered from several problems (Goldwyn et al., 2011). In a recent paper (Orio and Soudry, 2012) a more general method was derived, which had none of the previous problems, and was shown numerically to produce a very accurate approximation of the original Markov process description.

#### Derivation

According to Orio and Soudry (2012), if we define *x*^{(c)}_{k} to be the fraction of *c*-type channels in state *k*, and **x**^{(c)} to be a column vector composed of *x*^{(c)}_{k}, then

where ξ^{(c)} is a white noise vector process – meaning it has zero mean and auto-covariance

where **I** is the identity matrix, δ(*t*) is the Dirac delta function, and δ_{c, c′} = 1 if *c* = *c′* and 0 otherwise. Furthermore, **B**^{(c)} is defined so that in Equation (28) each component of ξ^{(c)}, which is associated with a transition pair *i* ⇋ *j*, is multiplied by $\sqrt{{\left(}{{A}}_{{i}{j}}^{{\left(}{c}{\right)}}{{x}}_{{j}}^{{\left(}{c}{\right)}}{+}{{A}}_{{j}{i}}^{{\left(}{c}{\right)}}{{x}}_{{i}}^{{\left(}{c}{\right)}}{\right)}{/}{{N}}^{{\left(}{c}{\right)}}}$, and appears in the equation for ${\dot{{x}}}_{{i}}^{{\left(}{c}{\right)}}$ and ${\dot{{x}}}_{{j}}^{{\left(}{c}{\right)}}$ with opposite signs. Note that **B**^{(c)} is not necessarily square since it has *K*^{(c)} rows but the number of columns is equal to the number of transition pairs.

We now need to combine Equation (28) for all *c* to obtain Equations (1–3). For simplicity, assume now that all ion channels types can be classified as either “rapid” or “slow” (this assumption can be relaxed). In this case we can concatenate all vectors related to rapid channels ${r}{\triangleq}{{\left(}{{x}}_{{\left(}{1}{\right)}}^{{\top}}{,}{\dots}{,}{{x}}_{{\left(}{R}{\right)}}^{{\top}}{\right)}}^{{\top}}$ and to slow channels ${s}{\triangleq}{{\left(}{{x}}_{{\left(}{R}{\text{\hspace{0.05em}}}{+}{\text{\hspace{0.05em}}}{1}{\right)}}^{{\top}}{,}{\dots}{,}{{x}}_{{\left(}{R}{+}{S}{\right)}}^{{\top}}{\right)}}^{{\top}}$, where *R* + *S* = *C*. We similarly define ξ_{r} and ξ_{s} together with the block matrices

and similarly for **B**_{r} and **B**_{s}. Note that **A**_{r} is square with size $\tilde{{M}}{=}{\displaystyle {{\sum}}_{{c}{\text{\hspace{0.05em}}}{=}{\text{\hspace{0.05em}}}{1}}^{{R}}{{K}}^{{\left(}{c}{\right)}}}$ rows while **A**_{s} is square with size $\tilde{{M}}{=}{\displaystyle {{\sum}}_{{c}{=}{R}{\text{\hspace{0.05em}}}{+}{\text{\hspace{0.05em}}}{1}}^{{R}{\text{\hspace{0.05em}}}{+}{\text{\hspace{0.05em}}}{S}}{{K}}^{{\left(}{c}{\right)}}}$ rows.

#### 4.1.1. Compressed formulation

In some cases, it is more convenient to re-write Equations (1–3) in a compressed form (this is always possible)

where **r, s**, and ξ_{r/s} have been redefined (their dimension has decreased), as we will show next. First, we comment that the main disadvantage is of these equations is that they are less compact and the notation is somewhat more cumbersome. However, there are also several advantages to this approach: (1) The vectors and matrices are smaller, (2) The rate and diffusion matrices do not have “troublesome” zero eigenvalues and can be diagonal (which is analytically convenient), (3) Most CBMs are written using this form (e.g., the HH model), so it is easier to apply our results using this formalism.

#### Derivation

To derive these compressed equations, we use the fact *x*^{(c)}_{k} denote fractions, so ∑_{k}*x*^{(c)}_{k} = 1, for all *c*. We can use this constraint, together with the irreducibility of the underlying ion channel process, to reduce by one the dimensionality of Equation (28) (see Soudry and Meir, 2012a for further details). Defining **I** to be the identity function, **J** to be the **I** with it last row removed, **e** ≜ (0, 0, …, 1)^{⊤}, **u** ≜ (1, 1, …, 1)^{⊤}, **G** ≜ (**I** − **eu**^{⊤}) **J**^{⊤}, ${\tilde{{A}}}^{{\left(}{c}{\right)}}{\triangleq}{J}{{A}}^{{\left(}{c}{\right)}}{G}$, ${\tilde{{B}}}^{{\left(}{c}{\right)}}{\triangleq}{J}{{B}}^{{\left(}{c}{\right)}}$ (with *x*^{(c)}_{K(c)} replaced by 1 − *x*_{1} − *x*_{2} … −*x*_{K(c)−1}) and **b**^{(c)} ≜ −**JA**^{(c)} **e** (${\tilde{{A}}}^{{\left(}{c}{\right)}}$ is invertible), we obtain the following equation for the reduced state vector **y**^{(c)} = **Jx**^{(c)} (which has only *K*^{(c)} − 1 states)

Again assuming that all channels can be classified as either “rapid” or “slow,” we concatenate all vectors related to rapid channels **r** ≜ (**y**^{⊤}_{(1)}, …, **y**^{⊤}_{(R)})^{⊤} and to slow channels **s** ≜ (**y**^{⊤}_{(R+1)}, …, **y**^{⊤}_{(R+S)})^{⊤}, where *R* + *S* = *C*. We obtain Equations (30, 31) by similarly defining **b**_{r}, **b**_{s}, ξ_{r} and ξ_{s} together with the block matrices

and similarly for ${\tilde{{B}}}_{{r}}$ and ${\tilde{{B}}}_{{s}}$. Note that ${\tilde{{A}}}_{{r}}$ is square with ${\tilde{{M}}}_{{r}}{=}{\displaystyle {{\sum}}_{{c}{\text{\hspace{0.05em}}}{=}{\text{\hspace{0.05em}}}{1}}^{{R}}{{K}}^{{\left(}{c}{\right)}}{-}{R}}$ rows while ${\tilde{{A}}}_{{s}}$ is square with ${\tilde{{M}}}_{{s}}{=}{\displaystyle {{\sum}}_{{c}{\text{\hspace{0.05em}}}{=}{\text{\hspace{0.05em}}}{R}{\text{\hspace{0.05em}}}{+}{\text{\hspace{0.05em}}}{1}}^{{R}{+}{S}}{{K}}^{{\left(}{c}{\right)}}{-}{S}}$ rows. Furthermore, it can be shown (Soudry and Meir, 2012a) that ${\tilde{{A}}}^{{\left(}{c}{\right)}}$ is a strictly stable matrix (all its eigenvalues are also eigenvalues of **A**^{(c)} except its zero eigenvalue, and so have a strictly negative real part), and ${\tilde{{D}}}^{{\left(}{c}{\right)}}{\triangleq}{\tilde{{B}}}^{{\left(}{c}{\right)}}{\tilde{{B}}}^{{\left(}{c}{\right)}{\top}}$ is positive definite (so all its eigenvalues are real and strictly positive). Therefore, also ${\tilde{{A}}}_{{r}}$ and ${\tilde{{A}}}_{{s}}$ are both strictly stable and ${\tilde{{D}}}_{{r}}$ and ${\tilde{{D}}}_{{s}}$ are positive definite. Therefore, if *V* is held constant, 〈**s**〉 and 〈**r**〉 tend to ${{s}}_{{\infty}}{=}{\tilde{{A}}}_{{s}}^{{-}{1}}{{b}}_{{s}}$ and ${{r}}_{{\infty}}{=}{\tilde{{A}}}_{{r}}^{{-}{1}}{{b}}_{{r}}$, respectively.

#### Example – the HHS model

The HHS model can be easily written using the compressed formulation. For example, comparing Equation (23) with Equation (31) we find that

Note that all the parameters are scalar in the HHS model, and so are not boldfaced, as in the general case.

### 4.2. Model Reduction

In this section we give additional technical details on section **2.2**. Specifically, we show how, given sparse spike stimulation and a few assumptions, it is possible to derive a simple reduced dynamical system (Equation 8) from the full equations of a general biophysical model for an isopotential neuron (Equations 1–3),

For more details on how its parameters and variables map to microscopic biophysical quantities, see section 4.1.

#### 4.2.1. The excitability constraint

As explained in section 2.1, we focus on models for **excitable** neurons describable by equations of the general form of Equations (36–38), rather than on arbitrary dynamical systems. This imposes some constraints on the parameters (Soudry and Meir, 2012b). Formally, recall that τ_{AP} and τ** _{s}** are the respective kinetic timescales of {

*V*,

**r**} and

**s**, and that τ

_{AP}< τ

**. Suppose we “freeze” the dynamics of**

_{s}**s**(so that effectively τ

_{s}= ∞) and allow only

*V*and

**r**to evolve in time. We say the original model describes an

**excitable**neuron, if the following conditions hold in this “semi-frozen” model:

- If
*I*(*t*) = 0, then for all initial conditions,*V*and**r**rapidly (within timescale τ_{AP}) relax to a constant and unique steady state (“rest”). - Assume that
*V*and**r**are near rest, and a short stimulation pulse is given with duration*t*_{stim}≤ τ_{AP}and amplitude*I*_{0}. For certain initial conditions and values of*I*_{0}, we get either a stereotypical “strong” response (“AP response”) or a stereotypical “weak” response in*V*(“no AP response”). Only for a very small set of initial conditions and values of*I*_{0}, do we get an “intermediate” response (“weak AP response”). By “stereotypical” we mean that the shape of response does not change much between trials or for different initial conditions in {*V*,**r**} (however, it can change with**s**).

Note that due to condition 1, such an excitable neuron is *not* oscillatory and does not spontaneously fire APs.

#### 4.2.2. Problem formulation

Formally, suppose an excitable neuron receives a train of *identical* stimuli, so

where ⊓(*x*) is a pulse, of width *t*_{w} (so ⊓(*x*) = 0 for *x* outside [0, *t*_{w}]). We denote by {*Y*_{m}}^{∞}_{m=−∞} the occurrence events of AP responses at times {*t*_{m}}^{∞}_{m=−∞}, i.e., immediately after each stimulation time *t*_{m} (Figure 1A),

Defining *T*_{m} ≜ *t*_{m + 1} − *t*_{m}, the inter-stimulus interval, and τ_{AP} as the upper timescale of an AP event (Figure 1B) we make the following assumption.

**Assumption 1**. *(a) The stimulation pulse width is small, t_{w} < τ_{AP}. (b) The spike times* {

*t*

_{m}}

^{∞}

_{m = 0}

*are temporally sparse, i.e*., τ

_{AP}≪

*T*

_{m}

*for every m (“no collisions”)*.

Our main objective here is to mathematically characterize the relation between {*Y*_{m}} and {*T*_{m}} under the most general conditions.

#### 4.2.3. Derivations

We define the sampled quantities *V*_{m} ≜ *V*(*t*_{m}), **r**_{m} ≜ **r**(*t*_{m}), **s**_{m} ≜ **s**(*t*_{m}), **x**_{m} ≜ (*V*_{m}, **r**^{⊤}_{m}, **s**^{⊤}_{m})^{⊤} and the history set _{m} ≜ ${\left\{}{{\left\{}{{x}}_{{k}}{\right\}}}_{{k}{=}{-}{\infty}}^{{m}}{,}{{\left\{}{{T}}_{{k}}{\right\}}}_{{k}{=}{-}{\infty}}^{{m}}{,}{{\left\{}{{Y}}_{{k}}{\right\}}}_{{k}{=}{-}{\infty}}^{{m}}{\right\}}$ (note that _{m} ⊂ _{m + 1}). The Stochastic Differential Equation (SDE) description in Equations (36–38) implies that **x**_{m} is a “state vector” with the *Markov property*, namely it is a sufficient statistic on the history to determine the probability of generating an AP at each stimulation,

and, together with *Y*_{m} and *T*_{m}, its own dynamics

which implies the following causality relations

This causality structure is reminiscent of the well known Hidden Markov Model (Rabiner, 1989), except that in the present case the output *Y*_{m}, affects the transition probability, and we have input *T*_{m}. Theoretically, if we knew the distributions in Equations (40) and (41), as well as the initial condition *P* (**x**_{0}), we could integrate and find an exact probabilistic I/O relation *P*({*Y*_{k}}^{m}_{k=0}|{*T*_{k}}^{m}_{k=0}. However, since it may be hard to find an expression for *P*(**x**_{m + 1}|**x**_{m}, *T*_{m}, *Y*_{m}) in general, we make a simplifying assumption.

**Assumption 2**. *T*_{m} ≪ *τ _{s} for every m*.

This assumption, together with Assumption 1 and the excitable nature of the CBM, renders the dynamics between stimulations relatively easy to understand. Specifically, between two consecutive stimulations, the fast variables (*V*(*t*), **r**(*t*)) follow stereotypically either the “AP response” (*Y*_{m} = 1) or the “no-AP response” (*Y*_{m} = 0), then equilibrate rapidly (within time τ_{AP}) to some quasi-stationary distribution *q*(*V*, **r**|**s**_{m}). Meanwhile, the slow variable **s** (*t*), starting from its initial condition at the time of the previous stimulation, changes slowly according to Equation (38), affected by the voltage trace of *V*(*t*) (through **A**_{s} (*V*)).

Summarizing this mathematically, we obtain the following approximations

Using these equations together with Equations (40) and (41), we obtain

Therefore, the “excitability” vector **s**_{m} can now replace the full state vector **x**_{m} = (*V*_{m}, **r**^{⊤}_{m}, **s**^{⊤}_{m})^{⊤} as the sufficient statistic that retains all relevant the information about the history of previous stimuli. Given the input {*T*_{m}}^{∞}_{m=−∞}, Equations (45) and (46) together completely specify a Markov process with the causality structure

Since the function *p*_{AP} (**s**) is not affected by the kinetics of **s**, it can be found by numerical simulation of a single AP using only Equations (36, 37), when **s** is held constant (see section 4.2.4). Now, instead of finding *P*(**s**_{m + 1}|**s**_{m}, *Y*_{m}, *T*_{m}) directly, we calculate the increments Δ**s**_{m} ≜ **s**_{m + 1} − **s**_{m} by integration of the SDE in Equation (38) between *t*_{m} and *t*_{m + 1}. First, we integrate the “predictable” part of the increment

to first order, where 〈*X*|*Y*〉 denotes the conditional expectation of *X* given *Y*. Note that **A**_{s} ~ *O*(τ_{s}^{−1}), so second order corrections are of order *O*((*T*_{m}τ_{s}^{−1})^{2}). Due to assumption 2, we have *T*_{m}τ_{s}^{−1} ≪ 1, so these corrections are negligible. Now,

where we defined

which are the average rates during rest, during an AP response and during a no-AP response, receptively. Note a similar notation was also used in Soudry and Meir (2012b) (e.g., Equations 2.15, 2.16 there), where the +/ − /0 were replaced with *H*/*M*/*L*.

Next, we calculate the remaining part of the increment, which is the “innovation,”

Obviously, 〈**n**_{m}|**s**_{m}, *T*_{m}, *Y*_{m}〉 = 0, and also

to first order. Note that **D**_{s} ~ *O*(τ_{s}^{−1}/*N*), where *N* = min_{c}*N*^{(c)}(*N*^{(c)} is the channel number of the *c*-type channel, as we defined in section 4.1), while Equation (53) has corrections of size *O*((*T*_{m}τ_{s}^{−1}/*N*)^{2}). Since *N* ≥ 1 (usually *N* ≫ 1), and due to assumption 2, we have *T*_{m}τ_{s}^{−1}/*N* ≪ 1, so these corrections are also negligible. Now,

where we defined

Additionally, we note that **A**_{±/0} (**s**_{m}) generally tend to be rather insensitive to changes in **s**_{m}. This is because the kinetic transition rates (which are used to construct **A**_{s} (*V*), as explained in section 4.1) tend to demonstrate this insensitivity when similarly averaged (see Figures 4B, 5 in Soudry and Meir, 2012b). The usual reasons behind this are (see appendix section B1 of Soudry and Meir, 2012b): (1) The common sigmoidal shape of the voltage dependency of the kinetic rates reduces their sensitivity to changes in the amplitude of the AP or the resting potential (2) The shape of the AP is relatively insensitive to **s** (3) The resting voltage is relatively insensitive to **s**. Therefore, in most cases we can approximate **A**_{±/0} (**s**_{m}) to be constant for simplicity (though this not critical to our subsequent results), as we shall henceforth do.

Additionally, we note that, strictly speaking, the voltage trace during an AP and at rest are stochastic, and therefore, **A**_{+}, **A**_{−}, **A**_{0}, **D**_{+}, **D**_{−} and **D**_{0} are stochastic. However, there are two factors that render this stochasticity negligible. First, the sigmoidal shape of the kinetic rates implies that **A**(*V*) is rather insensitive to fluctuations in the voltage (Figure 4 in our Soudry and Meir, 2012b). Second, noise mainly plays a role in the timing of AP initiation, but does not much affect the AP shape above threshold (see AP voltage traces in Schneidman et al., 1998, p. 1687). Therefore, we shall approximate **A**_{+}, **A**_{−}, **A**_{0}, **D**_{+}, **D**_{−} and **D**_{0} to be deterministic.

In summary, defining

and

we can write

with〈**n**_{m}|**s**_{m}, *T*_{m}, *Y*_{m}〉 = 0 and

These equations correspond to the result presented in Equation (8).

Finally, we note that the distribution of **n**_{m} given **s**_{m}, *T*_{m}, *Y*_{m} can be generally computed using the approach described in Orio and Soudry (2012). For example, it can be well approximated to have a normal distribution if channel numbers are sufficiently high and channel kinetics are not too slow (Orio and Soudry, 2012). In that case only knowledge of the variance (Equation 60) is sufficient to generate **n**_{m}. And so, using Equations (45), (59) and the full distribution of **n**_{m}, we can now simulate the neuronal response using a reduced model, more efficiently and concisely (with fewer parameters) than the full model (Equations 36, 38), since every time step is a stimulation event. The simulation time should shorten approximately by a factor of 〈*T*_{m}〉/*dt*, where *dt* is the full model simulation step. Note that the reduced model parameters, having been deduced from the full model itself, still retain a biophysical interpretation.

#### 4.2.4. Calculation of *p*_{AP} (**s**)

We numerically calculated *p*_{AP} (**s**) by disabling all the slow kinetics in the model – i.e., we only use Equations (1, 2) in main text, while **ṡ** = 0. Then, for every value of **s** we simulated this “semi-frozen” model numerically by first allowing **r** to relax to a steady state and then giving a stimulation pulse with amplitude *I*_{0}. We repeat this procedures 200 times for each **s**, and calculate *p*_{AP} (**s**) as the fraction of simulations that produced an AP. A few comments are in order: (1) In some cases (e.g., the HHMS model) we can use a shortcut and calculate *p*_{AP} (**s**) based on previous results. For example, suppose we know the probability function $\tilde{{p}}$_{AP} (*s*) for some model with a scalar ** s** and we make the substitution

*s*=

*h*(

**s**) where the components of

**s**represent independent and uncoupled channel types (Orio and Soudry, 2012) – then

*p*

_{AP}(

**s**) = $\tilde{{p}}$

_{AP}(

*h*(

**s**)) in the new model. (2) The timescale separation assumption τ

_{AP}≪

*T*

_{m}≪ τ

**implies that all the properties of the generated AP (amplitude, latency etc.) maintain similar causality relations with**

_{s}**s**

_{m}as does

*Y*

_{m}, so we can find their distribution using the same simulation we used to find

*p*

_{AP}(

**s**), similarly to the approach taken to compute

*L*(

**s**) in the deterministic setting (Soudry and Meir, 2012b). (3) Numerical results (Figure 6) suggest that we can generally write

where Φ is the cumulative distribution function of the normal distribution, *E* (**s**) is some “excitability function” (as defined in Soudry and Meir (2012b), so *p*_{AP} (**s**) = 0.5 on the threshold Θ = {**s**|*E* (**s**) = 0}), and *N*^{−1/2}_{r}, the “noisiness” of the rapid sub-system, directly affects the slope of *p*_{AP} (**s**) (Figure 6D, *bottom*). Also, as explained in Soudry and Meir (2012b), *E* (**s**) is usually monotonic in each component separately and increasing in *I*_{0} (Figure 6C, *top*) – which could be considered as just another component of ** s** which has zero rates.

#### 4.2.5. Compressed formulation – reduction

We can perform a very similar model reduction and linearization using the compressed formalism presented in section 4.1.1. We just need to define (or re-define) **A**_{±,0}, **b**_{±,0}, **D**_{±,0} (**s**_{m}), **A** (*Y*_{m}, *T*_{m}), **b** (*Y*_{m}, *T*_{m}) and **D** (*Y*_{m}, *T*_{m}, **s**_{m}) in the obvious way and repeat very similar derivations, arriving to

instead of Equation (59) (or Equation 8). Next, we demonstrate this for the HHS model.

#### 4.2.6. Example – HHS model reduction

We derive the parameters of the HHS reduced map. Recall that the HHS model is based on the compressed formulation. Following the reduction technique described in the previous sections, we numerically find the average rates γ_{±,0} and δ_{±,0} (as in Equations (2.15, 2.16) of Soudry and Meir (2012b), where there we denoted *H*/*M*/*L* instead of +/ − /0 here), τ_{AP} and *p*_{AP} (*s*) (section 4.2.4).

From Equations (32, 35), we find,

and so *A* (*Y*_{m}, *T*_{m}) and *D* (*Y*_{m}, *T*_{m}, *s*_{m}) are defined as in Equations (57) and (58), and similarly

We give for example some specific values: if τ_{AP} = 15 ms, then in the range *I*_{0} = 7.5−8.3 μ*A*, we have δ_{±,0} = 25.5−25.6 mHz, γ_{+} = 22.9−22.1 mHz, γ_{−} = 0.9−1.3μ Hz and γ_{0} = 0.29−0.28μ Hz.

Recall that these averaged kinetic rates are determined by the shape of the voltage dependent rates (γ (*V*) and δ (*V*), see Equation 125) (Soudry and Meir, 2012b). The relative values of the averaged kinetic rates determine what kind of information can be stored in *s* (which retains the “memory” of the neuron between stimulation). We qualitatively demonstrate this in Figure 7 depicting the values of γ_{±,0} for three different shapes of γ (*V*): when γ (*V*) is sigmoidal with high threshold, when it is sigmoidal with low threshold and when it is constant. These determine whether γ (*V*) is affected by the output (APs), the input (stimulation pulses) or neither. Therefore: (1) if γ (*V*) and δ (*V*) are independent of the voltage, then *s* cannot store any information on input or output. (2) if γ (*V*) or δ (*V*) have low voltage threshold, then *s* can directly store information on the input. (3) if γ (*V*) or δ (*V*) have high voltage threshold, then *s* can directly store information about the output. In the HHS model the inactivation rate γ has high threshold, while δ is voltage independent – therefore, *s* directly stores information on the output.

### 4.3. Linearization

In this section we present a more detailed account on how to arrive from the reduced model (mainly, Equation 8) to its linearized version (the results in Equations 11, 12).

First, we write the complete reduced model, using Equations (59), (60), and (45). The reduced model is a non-linear stochastic dynamic “state-space” system with *T*_{m}, the inter-stimulus interval lengths, serving as inputs, **s**_{m} representing the neuronal state, and *Y*_{m} the output. We have

where 〈**n**_{m}**n**^{⊤}_{m}|**s**_{m}, *T*_{m}, *Y*_{m}〉 = *T*_{m}**D** (*Y*_{m}, *T*_{m}, **s**_{m}),

and

and we defined

Based on the causality structure in Equation (42), it is straightforward to prove that *e*_{m} and **n**_{m} are uncorrelated white noise processes – i.e., 〈*e*_{m}〉 = 0, 〈**n**_{m}〉 = 〈*e*_{n}**n**_{m}〉 = **0** and 〈**n**_{m}**n**^{⊤}_{n}〉 = 〈**n**_{m}**n**^{⊤}_{m}〉 δ_{mn}, 〈*e*_{m}*e*_{n}〉 = 〈*e*^{2}_{m}〉 δ_{mn} where δ_{nm} = 1 if *n* = m and 0 otherwise.

We now examine the case where {*T*_{m}} is a Wide Sense Stationary (WSS) process (i.e., the first and second order statistics of the process are invariant to time shifts), with mean *T*_{*}, so that the assumptions τ_{AP} ≪ *T*_{m} ≪ τ** _{s}** are fulfilled with high probability. In this case the processes {

**s**

_{m}} and {

*Y*

_{m}} are also WSS, with constant means 〈

**s**

_{m}〉 =

**s**

_{*}and 〈

*Y*

_{m}〉 =

*p*

_{*}. Also, it is straightforward to verify that ${\langle}{\widehat{{T}}}_{{m}}{{n}}_{{n}}{\rangle}{=}{0}$, and ${\langle}{\widehat{{T}}}_{{m}}{{e}}_{{n}}{\rangle}{=}{0}$.

In order to linearize the system in Equations (59–66) we denote ${\widehat{{T}}}_{{m}}{\triangleq}{{T}}_{{m}}{-}{{T}}_{{\ast}}$, ${\widehat{{Y}}}_{{m}}{\triangleq}{{Y}}_{{m}}{-}{{p}}_{{\ast}}$, ${\widehat{{s}}}_{{m}}{\triangleq}{{s}}_{{m}}{-}{{s}}_{{\ast}}$, ${w}{\triangleq}{{\nabla}{{p}}_{{\text{AP}}}{|}}_{{{s}}^{{\ast}}}$. In order for this linearization to be accurate we require that ${\widehat{{s}}}_{{m}}$ is “small enough.”

**Assumption 3**. *With high probability*${\left|}{\widehat{{s}}}_{{m}}{\right|}{\ll}{\left|}{{s}}_{{\ast}}{\right|}$ *(component-wise) and*${\left|}{{w}}^{{\top}}{\widehat{{s}}}_{{m}}{\right|}{\gg}{\left|}{\widehat{{s}}}_{{m}}^{{\top}}{\left(}{{\nabla}{\nabla}{{p}}_{{\text{AP}}}{|}}_{{{s}}^{{\ast}}}{\right)}{\widehat{{s}}}_{{m}}{\right|}$.

This assumption essentially means that **s**_{*} = **s**_{*}(*p*_{*}, *T*_{*}) is a stable fixed point of the system (Equations 59–66), and stochastic fluctuations around it are small, compared to the size of the region {**s**|*p*_{AP} (**s**_{m}) ≠ 0, 1} (usually determined by the noise level of the rapid system {*V*, **r**}, see section 4.2.4). Note that the region is usually rather narrow (Figure 6) and therefore ${\left|}{\widehat{{s}}}_{{m}}{\right|}{\ll}{\left|}{{s}}_{{\ast}}{\right|}$ is often implied by this description. Given Assumption 3, we can approximate to first order

which allows us to linearize Equation (66). This essentially means that the components of ${\widehat{{s}}}_{{m}}$ determine the neuronal response linearly, with the components of **w** serving as the effective weights (related to the relevant conductances in the original full neuron model).

Next, we wish to linearize Equation (59). Using our assumptions, we obtain to first order

Taking expectations and using Equations (66) and (68), we obtain

to zeroth order. Defining the solution of this equation is **s**_{*}(*p*_{*}, *T*_{*}) and we can find *p*_{*} implicitly from

We write the explicit solution of this equation as *p*_{*}(*T*_{*}). Next, using ${\left|}{\widehat{{s}}}_{{m}}{\right|}{\ll}{\left|}{{s}}_{{\ast}}{\right|}$, Equation (70) and defining

we can approximate Equation (69) as

which, together with

yields a simple linear state space representation with ${\widehat{{T}}}_{{m}}$ as the input, ${\widehat{{s}}}_{{m}}$ as the state, ${\widehat{{Y}}}_{{m}}$ as the output and two uncorrelated white noise sources with variances

to first order.

#### 4.3.1. Derivation of w

From Equation (61), we note that generally we can write

where in many cases the excitability function *E* (**s**) has the form *E* (**s**) = *μ*^{⊤} **s** − θ, where the components of ** μ** are proportional to the relevant conductances (Soudry and Meir, 2012b). Therefore, if

then *E* (**s**) → ±∞, so in this case (assuming *E* (**s**_{*}) is not a particularly “pathological” function) we have

#### 4.3.2. Compressed formulation – linearization

In the compressed formulation (introduced in sections 4.1.1 and 4.2.5), we can perform similar linearization by re-defining **F** ≜ **I** + *T*_{*}**A** (*p*_{*}, *T*_{*}), **d** ≜ **A**_{0}**s**_{*} − **b**_{0}, **a** ≜ τ_{AP} ((**A**_{+} − **A**_{−})**s**_{*} − (**b**_{+} − **b**_{−})), and repeat very similar derivations, where now we can write more explicitly

instead of Equation (70).

#### 4.3.3. Example – HHS model linearization

Note again that all the parameters are scalar now, and so are not boldfaced, as in the general case. From Equations (71) and (81) we obtain **s**_{*} and *p*_{*} for a given *T*_{*}. Once *s*_{*} is known, from Equation (79) *w* can be obtained^{7}. Next, we denote the average inactivation rate at steady state by

and similarly for the recovery rate δ_{*}. And so, *s*_{*} = δ_{*}/(γ_{*} + δ_{*}), and

Denoting γ_{1} ≜ γ_{+} − γ_{−} and similarly for δ_{1}, we obtain

Finally, from Equations (77, 78) we find

### 4.4. Linear Systems Analysis

In section 2.4 we describe the neuronal dynamics using a linear system for the fluctuations, as depicted in Figure 1. This linear description allows us to use standard engineering tools to analyze the system. In this section we provide an easy to follow description on how this was done, for those unfamiliar with these topics.

#### 4.4.1. Second order statistics and linear systems

We start with a short reminder on some known results for stochastic processes (Papoulis and Pillai, 1965; Gardiner, 2004); these results are standard but are provided for completeness. These results will be used in later sections.

Assume {**x**_{m}} and {**y**_{m}} are two real-valued vector stochastic processes that are jointly wide-sense stationary (i.e., a simultaneous time shift of both processes will not change their first and second order statistics). We define the cross-covariance (recall that $\widehat{{x}}{=}{x}{-}{\langle}{x}{\rangle}$)

and the Cross-Power Spectral Density (CPSD), given by its Fourier transform

Additionally, the auto-covariance is defined as *R*_{x} ≜ *R*_{xx} and the corresponding Power Spectral Density (PSD) as *S*_{x} ≜ *S*_{xx}. Also, note that *R*_{yx}(*k*) = *R*_{xy}^{⊤}(−*k*) and so *S*_{yx}(*k*) = *S*_{xy}^{⊤}(−ω).

Suppose now that {**y**_{m}} is generated from a process {**x**_{m}} using a linear system: i.e., if the Fourier transform *x*(ω) ≜ ∑^{∞}_{k = −∞} *x*_{k}*e*^{−iωk} exists, then in the frequency domain

where **H**(ω) is a matrix-valued “transfer” function. Therefore, under some regularity conditions (allowing us to switch the order of integration end expectation),

And similarly

where in the second equality here we used an almost identical derivation as for *S*_{xy}(ω).

Note that if instead

where **x** and **z** are two uncorrelated signals, then we can write

where

Thus Equations (90) and (91), respectively give

#### 4.4.2. The second order statistics of our system

Previously, we derived Equations (11, 12), which describe the neuronal dynamics using a linear system, written in “state-space” form

where **n**_{m}, *e*_{m} and ${\widehat{{T}}}_{{m}}$ are uncorrelated, zero mean processes with the PSDs Σ_{n} ≜ *T*_{*}**D** (*p*_{*}, *T*_{*}, *s*_{*}), σ^{2}_{e} = *p*_{*} (1 − *p*_{*}) and *S*_{T}(ω), respectively.

In order to apply Equations (92) and (93) to our system we first need to find the transfer function of the system. Applying the Fourier transform to Equations (94, 95) gives

Re-arranging terms, we obtain

where we denoted

This gives the “closed loop” transfer functions of the system (including the effect of the feedback $\widehat{{Y}}{(}{\omega}{)}$). Next, combining Equations (98, 99) and Equations (92, 93) leads to explicit expressions for the PSDs and CPSDs.

For low frequencies it is sometimes more convenient to use the “continuous-time” versions of the PSDs, *S*_{xy}(*f*) ≜ *T*_{*}*S*_{xy}(ω)_{ω = 2π fT*} for *f* ≪ *T*^{−1}_{*}, which are approximated by

where

and we used the fact that **F** = **I** + *T*_{*}**A** (*p*_{*}, *T*_{*}) (Equation 72) and **Σ**_{n} = *T*_{*}**D** (*p*_{*}, *T*_{*}, **s**_{*}) (Equation 77).

Note that if the dimension of **s** is finite and there is no degeneracy, we can always write

where λ_{i}, the poles of *S*_{Y}(*f*), are determined solely by the poles of **H**_{c}(*f*) and *S*_{T}(*f*), while all the other parameters in Equation (106) affect only the constants *c*_{j}. Commonly, *S*_{T}(*f*) has no poles – for example, if *S*_{T}(*f*) is constant so *T*_{m} is a renewal process (e.g., the stimulation is periodic or Poisson). Therefore all poles of *S*_{Y}(*f*) (or the other PSDs) are determined by **H**_{c}(*f*), i.e., λ_{j} are the roots of the characteristic polynomial

#### 4.4.3. Spectral factorization

Equations (96) and (97) can be re-arranged as a *direct* I/O relation, formulated, for convenience, in the frequency domain (this can be either *f* or ω – in the section we use ω for brevity of notation, and *f* in other places). Specifically, this relation is of the form

so *v*_{m} = ^{−1} (*v*(ω)) is a single **scalar** “noise” process with zero mean and PSD σ^{2}_{v} (here ^{−1} is the inverse Fourier transform). This *v*_{m} process combines the contributions of *e*_{m} and **n**_{m}, which are the noise processes in the original system (in Equations 96, 97). Such a description, as in Equation (109), describes concisely the contributions of the input and noise to the output (an ARMAx model Lejung, 1999). Using 92 and 93 we respectively find that

Comparing Equation (102) with (110) we obtain

Comparing Equation (103) with (111), while using Equation (112), will yield the equation

This is a “spectral factorization” problem (Anderson and Moore, 1979), with solution

where

and

with **P** the solution of

derived from the general discrete-time algebraic Riccati equation. This can be verified by substitution

where in (1) we used the fact that **w**^{⊤} **H**_{c}(ω) = **w**^{⊤} **H**_{o}(ω) (1 − **w**^{⊤} **H**_{o}(ω) **a**)^{−1} from the Sherman–Morrison lemma, with **H**_{o}(ω) = (*e*^{iω} **I** − **F**)^{−1} being the “open loop” version of **H**_{c}(ω) (i.e., if **a** was zero), and in (2) we used the fact that **H**_{o}(ω) = *e*^{−iω}(**FH**_{o}(ω) + **I**).

#### 4.4.4. Optimal linear estimation of linear systems

Given that the neuronal dynamics are given by the linear system in Equations (96, 97), there are two different estimation problems one may be interested in. We may want to estimate, based on the history of the previous inputs and outputs ${{\left\{}{\widehat{{T}}}_{{k}}{,}{\widehat{{Y}}}_{{k}}{\right\}}}_{{k}{=}{-}{\infty}}^{{m}{\text{\hspace{0.05em}}}{-}{\text{\hspace{0.05em}}}{1}}$, either the *parameters* of the model (**F, w, a, d**, σ_{e} and Σ_{n}), or the *variables* in the model (${\widehat{{Y}}}_{{m}}$ or ${\widehat{{s}}}_{{m}}$). The first problem is generally termed a “system identification” problem (Lejung, 1999), while the second is a “filtering” (or prediction) problem (Anderson and Moore, 1979). Both are intimately related, and sometimes the solution of the second problem can yield a method of solving the first problem (e.g., section 3.3 in Anderson and Moore, 1979).

A relatively simple way to approach the second (filtering) problem involves the output decomposition we have found in section 4.4.3

Using this decomposition we can now write a new state-space representation for the system in terms of new state variable ${\widehat{{z}}}_{{m}}$,

which has the same output in the frequency domain (recall, from linear systems theory, that a single I/O relation can be generated by multiple state space realizations). This “innovation form” is particularly useful, since, given the entire history of the previous inputs and outputs ${{H}}_{{m}{\text{\hspace{0.05em}}}{-}{\text{\hspace{0.05em}}}{1}}{\triangleq}{{\left\{}{\widehat{{T}}}_{{k}}{,}{\widehat{{Y}}}_{{k}}{\right\}}}_{{k}{=}{-}{\infty}}^{{m}{\text{\hspace{0.05em}}}{-}{\text{\hspace{0.05em}}}{1}}$, we can recursively estimate the current state precisely (with zero error) (Anderson and Moore, 1979)

Given this precise estimate of ${\widehat{{z}}}_{{m}}$, the best linear estimate of ${\widehat{{Y}}}_{{m}}$ is simply

and the estimation error is simply

Since both the innovation form and the original form have the same second order statistics for the input–output, the optimal linear estimator (and its error) for ${\widehat{{Y}}}_{{m}}$ in the original system would be the same. Moreover, one can show (Anderson and Moore, 1979) that Equation (118) will also give the optimal linear estimate of ${\widehat{{s}}}_{{m}}$ in the original system, and with error **P** (Equation 117). This solution is the well-known “Kalman filter.”

#### 4.4.5. Example – HHS model power spectral densities

Substituting the parameters for the linearized map (Equations 85–89) into the expressions for the power-spectral densities (Equations 104–106), gives

Note that when *S*_{T}(*f*) ≡ 0 (i.e., periodical spike stimulus), *S*_{Y}(*f*) has the shape of high pass filter (Figure 3B, *top*). In contrast, *S*_{s}(*f*) (Figure 3B, *bottom*) and *S*_{YT}(*f*) both have the shape of a low pass filter (Figure 3D, *top*). From Equations (111) and (110) we know that *S*_{Y}(*f*) = |*H*^{int}(*f*)|^{2}σ^{2}_{v} and *S*_{YT}(*f*) = *H*^{ext}(*f*)*S*_{T}(*f*), respectively. Therefore, this indicates that *H*^{int}(*f*) and *H*^{ext}(*f*) are high pass and low pass filters, respectively.

#### 4.4.6. Power spectral densities of response features

So far we have concentrated on the PSD of the response *Y*_{m}. However, it is easy to extend our formalism to derive the PSDs of different features of the AP, such as its latency or amplitude. We exemplify this on the latency. In Soudry and Meir (2012b) we showed (Figure 3) that for deterministic CBMs, the latency of the AP generated in response to the *m*-th stimulation can be written as a function of the excitability *L*_{m} = *L* (**s**_{m}). In a stochastic model, we have instead

where ϕ_{m} is a zero mean, white noise process generated by the stochasticity of the rapid system. Since it is problematic to define the PSD of *L*_{m} if sometimes *Y*_{m} = 0, we focus on the case that *p*_{*} = 1, so we always have *Y*_{m} = 1. In this case, assuming again that the perturbations in ${\widehat{{s}}}_{{m}}$ are small, we can linearize

where **l** = ∇*L* (**s**)_{s = s*}, to obtain (using Equation 11)

where he **F** = **I** + *T*_{*}**A** (1, *T*_{*}). Therefore, it is straightforward to show that the PSD of the latency is

where σ^{2}_{ϕ} = 〈ϕ^{2}_{m}〉. Note that if latency is a good indicator of excitability, i.e., ** L (s)** changes similarly to

*p*(

**s**) so that

**l**∝

**w**, then

*S*

_{L}(

*f*) =

*c*

_{1}

*S*

_{Y}(

*f*) +

*c*

_{2}for some constants

*c*

_{1},

*c*

_{2}, when the input is periodic (

*T*

_{m}=

*T*

_{*}) and

*p*

_{*}→ 1.

### 4.5. Numerical Tests

MATLAB (2010b) code is available on the ModelDB website, with accession number 144993. In all the numerical simulations of the full stochastic Biophysical neuron model we used Equations (1–3) in main text. We used first order Euler–Maruyama integration with a time step of *dt* = 5 μs (quantitative results were verified also at *dt* = 0.5 μs). Each stimulation pulse was given as a square pulse with a width of *t*_{stim} = 0.5 ms and amplitude *I*_{0} (which were respectively named *t*_{0} and *I*_{0} in Soudry and Meir, 2012b). The results are not affected qualitatively by our choice of a square pulse shape. We define an AP to have occurred if, after the stimulation pulse was given, the measured voltage has crossed some threshold *V*_{th} (we use *V*_{th} = −10 mV in all cases). In all cases where direct stimulation is given, unless stated otherwise, we used periodic stimulation with *I*_{0} = 7.9 μA and *T*_{*} = 50 ms. Note that for the parameter values used, no APs are spontaneously generated, consistently with experimental results (Gal et al., 2010).

The PSDs were estimated using the Welch method and averaged over eight windows, unless 1/f behavior was observed, in which case we used a single window instead, since long term correlations may generate bias if averaging is used (Beran, 1994). Numerical estimation of the cross-PSD is more problematic. When estimating cross-spectra, estimation noise level can be quite high (proportional to the inverse coherence, according to Bendat and Piersol, 2000, p. 321). To estimate the level of estimation noise, we estimate the cross-spectrum with the input randomly shuffled (Figure 8). Since in this case there is no input–output correlation, this new estimate is pure noise. Finally, as suggested by the reviewer, we smoothed the resulting PSD (or cross-PSD) in all figures (except in Figure 8, where we aimed to show the level of estimation noise). To achieve uniform variance with low bias, we divided the spectrum into 30 logarithmically spaced segments from *f* = 10^{−3} Hz to the maximal frequency (*T*_{*}/2). In each segment *n* the PSD (or cross-PSD) was smoothed using a window of size *n*.

**Figure 8. Estimation noise in the cross-power spectral density**. To estimate the level of this noise in Figure 3D, we added ${{S}}_{{Y}\tilde{{T}}}{\left(}{f}{\right)}$ where ${\left\{}{\tilde{{T}}}_{{m}}{\right\}}$ is a shuffled version of {*T*_{m}}. Only when the estimated *S*_{YT}(*f*) is above ${{S}}_{{Y}\tilde{{T}}}{\left(}{f}{\right)}$, is its estimation valid. Therefore, in Figure 3D we show only this region (left of dashed black line), where estimation is valid.

Next, we describe the models used Figures 3–5 and provide their parameter values. These models have either been studied in the literature or are extensions of such models, which are meant to explore the limit for the validity of our analytic approximations. In all cases where direct stimulation is given, unless stated otherwise, we use periodic stimulation with *I*_{0} = 7.9 μA and *T*_{*} = 50 ms. Notice the form of the models is given in the (more popular) compressed formalism (section 4.1.1), which employs the normalization of state occupation probability to reduce the dimensionality of equations of Equations (2, 3) in the main text.

#### 4.5.1. The HHS model

The HHS model combines the Hodgkin–Huxley equations (Hodgkin and Huxley, 1952) with slow sodium inactivation (Chandler and Meves, 1970; Fleidervish et al., 1996). The model equations (Soudry and Meir, 2012b), which employ the uncoupled stochastic noise approximation, are

Most of the parameters are given their original values (as in Hodgkin and Huxley, 1952; Fleidervish et al., 1996):

where in all the rate functions *V* is used in units of mV. In order to obtain the specific spike shape and the latency transients observed in cortical neurons, some of the parameters were modified to

We emphasize that these specific choices do not affect any of our general arguments, but were chosen for consistency with experimental results (Gal et al., 2010). Estimates of channel number vary greatly (Soudry and Meir, 2012b). For simplicity, we chose *N* = *N*_{n} = *N*_{h} = *N*_{m} = *N*_{s}, and unless stated otherwise, we chose, by default *N* = 10^{6}, as in Soudry and Meir (2012b). Note that the HHS model is the same model presented in the paper with *M* = 1, ϕ_{s,1} = 1, *N*_{s,1} = *N, N*_{r, j} = *N* and ϕ_{r} = ϕ.

#### 4.5.2. The coupled HHS model

The coupled version of the HHS model uses the same parameters as the uncoupled version, and a similar voltage equation

where the variables *n*_{0} and *s*_{0}*m*_{0}*h*_{0} describe the respective fraction of potassium and sodium channels residing in the “open” state. To obtain the coupled model equations, we need to assume something about the structure of the ion channels. The original assumption by Hodgkin and Huxley was that the channel subunits (e.g., *m, n* and *h*) are independent. Over the years, it became apparent that this assumption is inaccurate, and the sodium channel kinetic subunits are, in fact, not independent (Ulbricht, 2005). However, it is not yet clear how the slow sodium inactivation is coupled to the rapid channel kinetics (e.g., Menon et al., 2009; Milescu et al., 2010), so we nevertheless used the original naive HH model assumption that the subunits are independent. In that case the potassium channel structure is given by (for brevity, the voltage dependence on the rates is henceforth ignored for this model)

while for the sodium channel it is described by

In this diagram, transition rates indicated between two boxed regions, imply that the same rates are used between all corresponding states in boxed regions. The corresponding 32 SDEs are derived using the method described in Orio and Soudry (2012) (or 30 equations if we use the compressed formalism). In this model we used *I*_{0} = 8.3 μA.

#### 4.5.3. The HHSTM model

In order to investigate the effect of a more “physiological” stimulation, we changed the HHS model and added synapses. We used the popular Tsodyks–Markram model for the effect of a synapse with short-term-depression on the somatic voltage (the model first appeared in Tsodyks and Markram (1997) and was slightly corrected in Tsodyks et al. (1998)). In the model *x, y* and *z* are the fractions of resources in the recovered, active and inactive states respectively, interacting through the system

Here the *z* → *x* rate is τ^{−1}_{rec}, the *x* → *y* rate is τ^{−1}_{in}, and the *x* → *y* rate is *U*_{SE}δ(*t* − *t*_{sp}), where δ(·) is the Dirac delta function, and *t*_{sp} is the pre-synaptic spike arrival time. The post-synaptic current is given by *I*_{s}(*t*) = *A*_{SE}*y*(*t*) where *A*_{SE} is a parameter. Additionally, we added noise to the model using the coupled SDE method (Orio and Soudry, 2012), assuming that the diagram in Equation (126), with the corresponding rates, hint at the underlying Markov kinetic structure, with *N* = 10^{6}. As in Figure 1B of Tsodyks and Markram (1997), we used τ_{in} = 3 ms, τ_{rec} = 800 ms and *U*_{SE} = 0.67. Additionally, we set *A*_{SE} = 70 μA to obtain an AP response in our model.

#### 4.5.4. The HHMS model

The HHMS model consists of many sodium currents, each with a different slow kinetic variable. The equations are identical to the HHS model, except that *g*_{Na}*s* is replaced by *g*_{Na}*M*^{−1} ∑ ^{M}_{i=1} *s*_{i}, where *s*_{1} has the same equation as *s* in the HHS model, and for *i* > 2,

with ϕ_{s, i} = ϵ^{i} and *N*_{s, i} = *N*_{0}ϵ^{iη}, where γ (*V*) and δ (*V*) are taken from the HHS model. Unless mentioned otherwise, we chose as default ϵ = 0.2, η = 1.5, *M* = 5, and *N*_{0} = *N* as in Figure 5.

#### 4.5.5. The multiplicative hhms model

The Multiplicative HHMS model is identical to the HHMS model with η = 1, except that ${\overline{{g}}}_{{N}{a}}{{M}}^{{-}{1}}{\displaystyle {{\sum}}_{{i}{\text{\hspace{0.05em}}}{=}{\text{\hspace{0.05em}}}{1}}^{{M}}{{s}}_{{i}}}$ is replaced with ${\overline{{g}}}_{{N}{a}}{\displaystyle {{\prod}}_{{i}{\text{\hspace{0.05em}}}{=}{\text{\hspace{0.05em}}}{1}}^{{M}}{{s}}_{{i}}}$.

#### 4.5.6. The HHSIP model

The HHSIP model equations (Soudry and Meir, 2012b) are identical to the HHS model equations, except that *s* is renamed to *s*_{1} and an Inactivating Potassium current was added to the voltage equation, where

with *g*_{M} = 0.05*g*_{K} and

where *N*_{s2} = *N* and

Again, in all the rate functions *V* is used in mV units. In this model we used *I*_{0} = 8.3 μA and *T*_{*} = 33 ms.

#### 4.5.7. The HHMSIP model

The HHMSIP model combines HHSIP and HHMS. Its equations are identical to the HHMS model with η = 2, except they also contain the *I*_{K} current from the HHSIP model. In this model we used *I*_{0} = 8.3 μA and *T*_{*} = 33 ms, unless otherwise specified.

## Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

## Acknowledgments

The authors are grateful to O. Barak, N. Brenner, Y. Elhanati, A. Gal, T. Knafo, Y. Kafri, S. Marom, and J. Schiller for insightful discussions and for reviewing parts of this manuscript. The research was partially funded by the Technion V.P.R. fund and by the Intel Collaborative Research Institute for Computational Intelligence (ICRI-CI).

## Footnotes

1. ^A semi-analytic derivation is an analytic derivation in which some terms are obtained by relatively simple numerics. See 2.2 for our implementation.

2. ^We demonstrated that such noise should strongly affect the neuronal response to sparse stimulation (Soudry and Meir, 2012b).

3. ^I.e., if ∀*t*:*I*(*t*) = 0, then the probability that a neuron will fire is negligible – on any relevant finite time interval (e.g., minutes or days).

4. ^E.g., as in Equations (50–52). Note also a similar notation was also used in Soudry and Meir (2012b) (e.g., Equations 2.15, 2.16), where we used *H*/*M*/*L* instead of +/−/0.

5. ^Later we shall demonstrate numerically that this is not a necessary condition.

6. ^Note a more general Box–Jenkins model is not required, since the poles of *H*^{ext}(*f*) and *H*^{int}(*f*) are identical (assuming no pole-zero cancelation).

7. ^Also, as explained in section 4.3.1, we approximately have *w* ∝ *g*_{Na}, from Equation (21).

## References

Anderson, B. D. O., and Moore, J. B. (1979). *Optimal Filtering*, Vol. 11. Englewood Cliffs, NJ: Prentice Hall.

Bean, B. P. (2007). The action potential in mammalian central neurons. *Nat. Rev. Neurosci*. 8, 451–465. doi: 10.1038/nrn2148

Bendat, J. S., and Piersol, A. G. (2000). *Random Data Analysis and Measurement Procedures*, Vol. 11, 3rd edn. New York, NY: Wiley.

Beran, J. (1992). A goodness-of-fit test for time series with long range dependence. *J. R. Stat. Soc. Ser. B* 54, 749–760.

Brecht, M., Schneider, M., Sakmann, B., and Margrie, T. W. (2004). Whisker movements evoked by stimulation of single pyramidal cells in rat motor cortex. *Nature* 427, 704–710. doi: 10.1038/nature02266

Chandler, W. K., and Meves, H. (1970). Slow changes in membrane permeability and long-lasting action potentials in axons perfused with fluoride solutions. *J. Physiol*. 211, 707–728.

Channelpedia. Available online at: http://channelpedia.epfl.ch/

Colquhoun, D., and Hawkes, A. G. (1981). On the stochastic properties of single ion channels. *Proc. R. Soc. Lond. Ser. B Biol. Sci*. 211, 205–235. doi: 10.1098/rspb.1981.0003

Contou-Carrere, M. N. (2011). Model reduction of multi-scale chemical Langevin equations. *Syst. Control Lett*. 60, 75–86. doi: 10.1016/j.sysconle.2010.10.011

De Col, R., Messlinger, K., and Carr, R. W. (2008). Conduction velocity is regulated by sodium channel inactivation in unmyelinated axons innervating the rat cranial meninges. *J. Physiol*. 586, 1089–1103. doi: 10.1113/jphysiol.2007.145383

De Paola, V., Holtmaat, A., Knott, G., Song, S., Wilbrecht, L., Caroni, P., et al. (2006). Cell type-specific structural plasticity of axonal branches and boutons in the adult neocortex. *Neuron* 49, 861–875. doi: 10.1016/j.neuron.2006.02.017

Debanne, D., Campanac, E., Bialowas, A., and Carlier, E. (2011). Axon physiology. *Physiol. Rev*. 91, 555–602. doi: 10.1152/physrev.00048.2009

Druckmann, S., Berger, T. K., Schürmann, F., Hill, S., Markram, H., and Segev, I. (2011). Effective stimuli for constructing reliable neuron models. *PLoS Comput. Biol*. 7:e1002133. doi: 10.1371/journal.pcbi.1002133

Elul, R., and Adey, W. R. (1966). Instability of firing threshold and “remote” activation in cortical neurons. *Nature* 212, 1424–1425. doi: 10.1038/2121424a0

Ermentrout, B., and Terman, D. (2010). *Mathematical Foundations of Neuroscience*, Vol. 35. New York, NY: Springer. doi: 10.1007/978-0-387-87708-2

Fleidervish, I. A., Friedman, A., and Gutnick, M. J. (1996). Slow inactivation of Na+ current and slow cumulative spike adaptation in mouse and guinea-pig neocortical neurones in slices. *J. Physiol*. 493, 83–97.

Fox, R. F., and Lu, Y. N. (1994). Emergent collective behavior in large numbers of globally coupled independently stochastic ion channels. *Phys. Rev. E* 49, 3421–3431. doi: 10.1103/PhysRevE.49.3421

Gal, A., Eytan, D., Wallach, A., Sandler, M., Schiller, J., and Marom, S. (2010). Dynamics of excitability over extended timescales in cultured cortical neurons. *J. Neurosci*. 30, 16332–16342. doi: 10.1523/JNEUROSCI.4859-10.2010

Gal, A., and Marom, S. (2013). Entrainment of the intrinsic dynamics of single isolated neurons by natural-like input. *J. Neurosci*. 33, 7912–7918. doi: 10.1523/JNEUROSCI.3763-12.2013

Gardiner, C. W. (2004). *Handbook of Stochastic Methods*, 3rd edn. Berlin: Springer-Verlag. doi: 10.1007/978-3-662-05389-8

Gerstner, W., and Naud, R. (2009). How good are neuron models? *Science* 326, 379–380. doi: 10.1126/science.1181936

Goldwyn, J. H., Imennov, N. S., Famulare, M., and Shea-Brown, E. (2011). Stochastic differential equation models for ion channel noise in Hodgkin-Huxley neurons. *Phys. Rev. E* 83, 041908. doi: 10.1103/PhysRevE.83.041908

Goldwyn, J. H., Rubinstein, J. T., and Shea-Brown, E. (2012). A point process framework for modeling electrical stimulation of the auditory nerve. *J. Neurophysiol*. 108, 1430–1452. doi: 10.1152/jn.00095.2012

Grubb, M. S., and Burrone, J. (2010). Activity-dependent relocation of the axon initial segment fine-tunes neuronal excitability. *Nature* 465, 1070–1074. doi: 10.1038/nature09160

Hodgkin, A. L., and Huxley, A. F. (1952). A quantitative description of membrane current and its application to conduction and excitation in nerve. *J. Physiol*. 117, 500.

Huys, Q. J. M., Ahrens, M. B., and Paninski, L. (2006). Efficient estimation of detailed single-neuron models. *J. Neurophysiol*. 96, 872. doi: 10.1152/jn.00079.2006

Ikegaya, Y, Sasaki, T., Ishikawa, D., Honma, N., Tao, K., Takahashi, N., et al. (2012). Interpyramid spike transmission stabilizes the sparseness of recurrent network activity. *Cereb. Cortex* 23, 293–304. doi: 10.1093/cercor/bhs006

Jugloff, D. G. M. (2000). Internalization of the Kv1.4 potassium channel is suppressed by clustering interactions with PSD-95. *J. Biol. Chem*. 275, 1357–1364. doi: 10.1074/jbc.275.2.1357

Kaplan, D. T., Clay, J. R., Manning, T., Glass, L., Guevara, M. R., and Shrier, A. (1996). Subthreshold dynamics in periodically stimulated squid giant axons. *Phys. Rev. Lett*. 76, 4074–4077. doi: 10.1103/PhysRevLett.76.4074

Kasischke, K. A., Vishwasrao, H. D. D., Fisher, P. J., Zipfel, W. R., and Webb, W. W. (2004). Neural activity triggers neuronal oxidative metabolism followed by astrocytic glycolysis. *Science (New York, N.Y.)* 305, 99–103. doi: 10.1126/science.1096485

Koch, C., and Segev, I. (1989). *Methods in Neuronal Modeling: From Ions to Networks*, Vol. 484, 2nd edn. Cambridge: MIT Press.

Komlósi, G., Molnár, G., Rózsa, M., Oláh, S., Barzó, P., and Tamás, G. (2012). Fluoxetine (prozac) and serotonin act on excitatory synaptic transmission to suppress single layer 2/3 pyramidal neuron-triggered cell assemblies in the human prefrontal cortex. *J. Neurosci*. 32, 16369–16378. doi: 10.1523/JNEUROSCI.2618-12.2012

Lejung, L. (1999). *System Identification: Theory for the User*. Upper Saddle River, NJ: PTR Prentice Hall.

Levitan, I. B. (1994). Modulation of ion channels by protein phosphorylation and dephosphorylation. *Annu. Rev. Physiol*. 11, 193–212. doi: 10.1146/annurev.ph.56.030194.001205

Li, C. Y. T., Poo, M. M., and Y D. (2009). Burst spiking of a single cortical neuron modifies global brain state. *Science* 324, 643–646. doi: 10.1126/science.1169957

Linaro, D., Storace, M., and Mattia, M. (2011). Inferring network dynamics and neuron properties from population recordings. *Front. Comput. Neurosci*. 5:43. doi: 10.3389/fncom.2011.00043

Marom, S. (2010). Neural timescales or lack thereof. *Prog. Neurobiol*. 90, 16–28. doi: 10.1016/j.pneurobio.2009.10.003

Menon, V., Spruston, N., and Kath, W. L. (2009). A state-mutating genetic algorithm to design ion-channel models. *Proc. Natl. Acad. Sci. U.S.A*. 106, 16829–16834. doi: 10.1073/pnas.0903766106

Migliore, M., Cannia, C., Lytton, W. W., Markram, H., and Hines, M. L. (2006). Parallel network simulations with NEURON. *J. Comput. Neurosci*. 21, 119–129. doi: 10.1007/s10827-006-7949-5

Milescu, L. S., Yamanishi, T., Ptak, K., and Smith, J. C. (2010). Kinetic properties and functional dynamics of sodium channels during repetitive spiking in a slow pacemaker neuron. *J. Neurosci*. 30, 12113–12127. doi: 10.1523/JNEUROSCI.0445-10.2010

Molnár, G., Oláh, S., Komlósi, G., Füle, M., Szabadics, J., Varga, C., et al. (2008). Complex events initiated by individual spikes in the human cerebral cortex. *PLoS Biol*. 6:e222. doi: 10.1371/journal.pbio.0060222

Monjaraz, E., Navarrete, A., Lopez-Santiago, L. F., Vega, A. V., and Cota, G. (2000). L-type calcium channel activity regulates sodium channel levels in rat pituitary GH3 cells. *J. Physiol*. 523, 45–55. doi: 10.1111/j.1469-7793.2000.00045.x

Neher, E., and Sakmann, B. (1976). Single-channel currents recorded from membrane of denervated frog muscle fibres. *Nature* 260, 799–802. doi: 10.1038/260799a0

Nishiyama, H., Fukaya, M., and Watanabe, M. (2007). Axonal motility and its modulation by activity are branch-type specific in the intact adult cerebellum. *Neuron* 56, 472–487. doi: 10.1016/j.neuron.2007.09.010

Orio, P., and Soudry, D. (2012). Simple, fast and accurate implementation of the diffusion approximation algorithm for stochastic ion channels with multiple states. *PLoS ONE* 7:e36670. doi: 10.1371/journal.pone.0036670

Papoulis, A., and Pillai, S. U. (1965). *Probability, Random Variables, and Stochastic Processes*. New York, NY: McGraw-Hill.

Rabiner, L. R. (1989). A tutorial on hidden Markov models and selected applications in speech recognition. *Proc. IEEE* 77, 257–286. doi: 10.1109/5.18626

Roth, A., and Häusser, M. (2001). Compartmental models of rat cerebellar Purkinje cells based on simultaneous somatic and dendritic patch-clamp recordings. *J. Physiol*. 535, 445–472. doi: 10.1111/j.1469-7793.2001.00445.x

Schneidman, E., Freedman, B., and Segev, I. (1998). Ion channel stochasticity may be critical in determining the reliability and precision of spike timing. *Neural Comput*. 10, 1679–1703. doi: 10.1162/089976698300017089

Silver, I. A., Deas, J., and Erecinska, M. (1997). Ion homeostasis in brain cells: differences in intracellular ion responses to energy limitation between cultured neurons and glial cells. *Neuroscience* 78, 589–601. doi: 10.1016/S0306-4522(96)00600-8

Sjöström, P. J. J., Rancz, E. A. A., Roth, A., and Häusser, M. (2008). Dendritic excitability and synaptic plasticity. *Physiol. Rev*. 88, 769–840. doi: 10.1152/physrev.00016.2007

Song, S., Sjöström, P. J. J., Reigl, M., Nelson, S. B., and Chklovskii, D. B. (2005). Highly nonrandom features of synaptic connectivity in local cortical circuits. *PLoS Biol*. 3:e68. doi: 10.1371/journal.pbio.0030068

Soudry, D., and Meir, R. (2012a). An exact reduction of the master equation to a strictly stable system with an explicit expression for the stationary distribution. arXiv:1207.4436.

Soudry, D., and Meir, R. (2012b). Conductance-based neuron models and the slow dynamics of excitability. *Front. Comput. Neurosci*. 6:4. doi: 10.3389/fncom.2012.00004

Soudry, D., and Meir, R. (2014). The neuronal response at extended timescales: long term correlations without long memory. *Front. Comput. Neurosci*. 8:35. doi: 10.3389/fncom.2014.00035

Staub, O., Gautschi, I., Ishikawa, T., Breitschopf, K., Ciechanover, A., Schild, L., et al. (1997). Regulation of stability and function of the epithelial Na+ channel (ENaC) by ubiquitination. *EMBO J*. 16, 6325–6336. doi: 10.1093/emboj/16.21.6325

Toib, A., Lyakhov, V., and Marom, S. (1998). Interaction between duration of activity and time course of recovery from slow inactivation in mammalian brain Na+ channels. *J. Neurosci*. 18, 1893–1903.

Tsodyks, M., and Markram, H. (1997). The neural code between neocortical pyramidal neurons depends on neurotransmitter release probability. *Proc. Natl. Acad. Sci. U.S.A*. 94, 719–723. doi: 10.1073/pnas.94.2.719

Tsodyks, M., Pawelzik, K., and Markram, H. (1998). Neural networks with dynamic synapses. *Neural Comput*. 10, 821–835. doi: 10.1162/089976698300017502

Ulbricht, W. (2005). Sodium channel inactivation: molecular determinants and modulation. *Physiol. Rev*. 85, 1271–1301. doi: 10.1152/physrev.00024.2004

Keywords: conductance based neuron models, noise, ion channels, adaptation, power spectral density, linear response, system identification, analytical methods

Citation: Soudry D and Meir R (2014) The neuronal response at extended timescales: a linearized spiking input–output relation. *Front. Comput. Neurosci*. **8**:29. doi: 10.3389/fncom.2014.00029

Received: 20 December 2013; Accepted: 24 February 2014;

Published online: 02 April 2014.

Edited by:

David Hansel, University of Paris, FranceReviewed by:

Maurizio Mattia, Istituto Superiore di Sanità, ItalyJoaquín J. Torres, University of Granada, Spain

Copyright © 2014 Soudry and Meir. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Daniel Soudry, Department of Statistics, Center for Theoretical Neuroscience, Columbia University, 1255 Amsterdam Avenue, New York, NY 10027, USA e-mail: daniel.soudry@gmail.com