Fokker–Planck and Fortet Equation-Based Parameter Estimation for a Leaky Integrate-and-Fire Model with Sinusoidal and Stochastic Forcing
- Alexandre Iolov^{1, 2}Email author,
- Susanne Ditlevsen^{2} and
- André Longtin^{3}
https://doi.org/10.1186/2190-8567-4-4
© A. Iolov et al.; licensee Springer 2014
Received: 30 December 2012
Accepted: 6 June 2013
Published: 17 April 2014
Abstract
Analysis of sinusoidal noisy leaky integrate-and-fire models and comparison with experimental data are important to understand the neural code and neural synchronization and rhythms. In this paper, we propose two methods to estimate input parameters using interspike interval data only. One is based on numerical solutions of the Fokker–Planck equation, and the other is based on an integral equation, which is fulfilled by the interspike interval probability density. This generalizes previous methods tailored to stationary data to the case of time-dependent input. The main contribution is a binning method to circumvent the problems of nonstationarity, and an easy-to-implement initializer for the numerical procedures. The methods are compared on simulated data.
List of Abbreviations
LIF: Leaky integrate-and-fire
ISI: Interspike interval
SDE: Stochastic differential equation
PDE: Partial differential equation
Keywords
First-passage times Stochastic neuron models Parameter estimation from stopping times Fortet integral equation Fokker–Planck equation1 Introduction
Information processing in the nervous system is carried out by spike timings in neurons. To study the neural code in such a complicated system, a first step is to understand signal processing and transmission in single neurons. Stochastic leaky integrate-and-fire (LIF) neuronal models are a good compromise between biophysical realism and mathematical tractability, and are commonly applied as theoretical tools to study properties of real neuronal systems. A central issue is then to perform statistical inference from experimental data and estimate model parameters. Many electrophysiological experiments on neurons, namely extra-cellular recordings, are only capable of detecting the time of the spike and not the detailed voltage trajectory leading up to the spike. Estimating the parameters of the LIF model from this type of data is equivalent to estimating the parameters of a stochastic model from the statistics of the first-passage times only. A common assumption is that the data are well described by a renewal process, thus basing the statistical inference on the interspike intervals (ISIs), assuming these are realizations of independent and identically distributed random variables. Since only partial information about the process is available, the statistical problem becomes more difficult, and no explicit expression for the likelihood is available.
Different methods have been proposed. In the seminal paper [1], a point process approach is proposed. The spike trains of a collection of neurons are represented as counting processes. Time is discretized and the point processes approximated by 0–1 time series. Then the probability of firing in the next time interval is modeled as a function of the spike history. In this way, maximum likelihood estimation is feasible. External stimuli are not considered. In [2], a numerically involved moment method is developed. It uses the first two moments of the first-passage times of the Ornstein–Uhlenbeck process to a constant threshold, which are given as series expressions, and equates them to their empirical counterparts. In [3, 4], certain explicit moment relations derived from the Laplace transform of the first-passage time distribution are applied, but these are only valid under stimulation (supra-threshold regime). In [5], inference is based on numerical inversion of the Laplace transform. In [6], a functional of a three-dimensional Bessel bridge is applied to obtain a maximum likelihood estimator. None of these methods are feasible to extend to the time-inhomogeneous case, which is of our interest. In [7, 8], an integral equation is used to derive an estimator in the time-homogeneous setting. This approach is readily extended to time varying input, which we will explore in this paper. Some of the above methods are compared in [9]. Finally, a review of estimation methods is provided in [10].
Many sensory stimuli, like sound, contain an oscillatory component [11, 12]. Such inputs will cause oscillating membrane potentials in the neuron, generating rhythmic spiking patterns. The oscillation frequency determines the basic rhythm of spiking, and is considered to be significant for neuronal information processing. The dynamics of periodically forced neuron models have been extensively studied; see [13–18] and references therein. Even so, attempts to solve the estimation problem in these nonstationary settings have been rare. One problem is that the ISIs are no longer independent nor identically distributed. In [19], a more complicated model with linear filters is considered, allowing also for the spike history to influence the membrane potential dynamics. The estimation problem is solved through numerical solutions to the Fokker–Planck equation, and it is shown that the log-likelihood is concave, thus ensuring a global maximum; see also [20, 21]. Because their model is more involved, some approximations to the solution of the Fokker–Planck equation are applied, to ensure acceptable computing times. We will apply the full Fokker–Planck equation to solve our estimation problem, since the computing time is always lower than 2 seconds for a sample size of 1000 spikes.
In this paper, we thus describe and discuss two methods to estimate parameters of LIF models with the added complexity of a time-varying input current. We assume that the time-varying current is a sinusoidal wave, but we believe that the approaches generalize to an arbitrary periodic forcing with known frequency. One approach relies on the Fortet integral equation, which is readily extended to the time-inhomogeneous case. An advantage of this approach is that if the transition density of the diffusion in the LIF model is known, as is the case for the Ornstein–Uhlenbeck and the Feller model, the computational burden is limited. A second approach involves numerical solution of the Fokker–Planck equation, where the time-dependence is explicitly accounted for. After a numerical differentiation, the likelihood function can be calculated providing the maximum likelihood estimator. Nevertheless, we chose an alternative loss function which seems marginally more robust, directly comparing the survival function provided by the solution of the Fokker–Planck equation with its empirical counterpart. The two approaches give similar results and they are more carefully compared in the supplementary online material. Both methods need sensible starting values for the optimization algorithms, and we provide an easy-to-implement initializer. The estimation procedures are compared on simulated data and we find that both algorithms are able to find estimates close to the true values for several different dynamical regimes. We find that for small sample sizes the Fokker–Planck algorithm can be considered marginally preferable, whereas for larger sample sizes the Fortet algorithm becomes marginally superior. Moreover, at high frequencies of the sinusoidal forcing, the Fortet is better at identifying the parameters, though in general there is less information in the data to distinguish between a constant input and the amplitude of the periodic forcing.
2 Model
Here, μ is a bias current acting on the cell, τ is the decay time, A and ω are the amplitude and (angular) frequency of the sinusoidal current acting on the cell, σ is the strength of the stochastic fluctuations, $W={\{{W}_{t}\}}_{t\ge 0}$ is a standard Wiener process, and ${t}_{n}^{+}$ denotes the right limit taken at ${t}_{n}$. A spike occurs when the membrane voltage $V(t)$ crosses a voltage threshold, ${v}_{\mathrm{th}}$, and then $V(t)$ is instantaneously reset to the resting potential ${v}_{0}$. The difference between subsequent spike times, ${J}_{n}={t}_{n}-{t}_{n-1}$, is called the interspike interval (ISI).
see the discussion in [14], which can be directly inferred from the solution in Eq. (12) below. In both the supra-threshold and supersinusoidal regimes, $\alpha +\gamma /\sqrt{1+{\Omega}^{2}}>1$. The difference between the two is that in the supra-threshold regime the constant bias current alone is sufficient for spikes to occur, also in absence of noise, that is, $\alpha >1$. In the supersinusoidal regime, the sinusoidal current is necessary for spikes to occur in absence of noise, that is, $\alpha +\gamma /\sqrt{1+{\Omega}^{2}}>1$ and $\alpha \le 1$. In the critical regime, the sum of the two terms is just barely enough to guarantee deterministic spiking, that is $\alpha +\gamma /\sqrt{1+{\Omega}^{2}}\approx 1$. Finally, in the subthreshold regime, there would be no spikes without the noise, $\alpha +\gamma /\sqrt{1+{\Omega}^{2}}<1$.
Example of α, β, γ parameter values for the different regimes, given $\Omega =1$
Regime name | α | β | γ |
---|---|---|---|
Supra-threshold | 1.40 | 0.30 | 0.14 |
Supersinusoidal | 0.10 | 0.30 | 1.98 |
Critical | 0.50 | 0.30 | 0.71 |
Subthreshold | 0.40 | 0.30 | 0.57 |
With regards to Figs. 2 and 3, it is worth noting explicitly that combinations of noise and sinusoidal forcing can cause firing patterns in which spikes are phase locked, but skip a certain number of cycles. This leads to multimodal ISI densities. There are many different dynamical mechanisms that can yield such patterns, and the particular correlations between the ISIs will depend on the underlying voltage dynamics (which, in our case, we assume to be given by Eq. (1)); in particular, it may be difficult to distinguish whether the dynamics are subthreshold or supra-threshold, since both can show similar ISI densities; see [22].
2.1 Basic ISI Probability Density Functions
The subscript ϕ is to stress that g, G, and $\overline{G}$ depend on the value of ${\varphi}_{n}$ in Eq. (3). This is the formal statement that in a sinusoidally-driven neuron, the interspike intervals are not identically distributed, and are only independent conditioned on the sinusoidal phase at an interval’s onset. Knowing these distributions would provide the likelihood function, offering estimation by the preferred method of choice, the maximum likelihood estimator. Unfortunately, explicit expressions for the ISI distribution are not available except for the special case of $\gamma =0$ and $\alpha =1$; see [3]. Different representations of the likelihood function are available though, see [23], one of which we will use below.
2.2 Fokker–Planck Equation with Absorbing Boundaries
Equation (7) forms the basis of one of the methods below for estimating the structural parameters from the observed data.
This choice requires some explanation. In the $t\to \mathrm{\infty}$ limit, the distribution of ${X}_{t}$ in Eq. (3) without thresholding is Gaussian with mean given by Eq. (12) (below) and variance equal to ${\beta}^{2}/2$. Thus, to truncate the computational domain for the thresholded process from below, we take the lowest value of the asymptotic mean, $\alpha -\gamma /\sqrt{1+{\Omega}^{2}}$, then from this we subtract two standard deviations, $2\beta /\sqrt{2}$ and set the result to be the lower bound, ${x}_{-}$. Finally, if this value for ${x}_{-}$ happens to be larger than −0.25, we enforce that ${x}_{-}\le -0.25$.
Numerical considerations lead us to solve for F, instead of f, since delta functions are difficult to represent in floating point, while the initial conditions for F, the Heaviside step function, $H(x)$, faces no such difficulties [24]. The Heaviside step function is defined to be equal to 0 for $x<0$ and to be equal to 1 for $x\ge 0$. At this point, we need to derive the PDE for the distribution F, starting from the PDE for the density, f, Eq. (5).
The right-hand side in Eq. (8) is precisely the reflecting boundary condition on f once we recall that ${\partial}_{x}F=f$. Therefore, $C(t)\equiv 0$.
2.3 Fortet Equation
The left-hand side is simply the probability of exceeding ${v}_{\mathrm{th}}$ at time t starting at ${v}_{0}$ at time 0. This can also be written as the probability of hitting ${v}_{\mathrm{th}}$ for the first time at time $\tau <t$ and then exceeding ${v}_{\mathrm{th}}$ at time t starting at ${v}_{\mathrm{th}}$ at time τ, integrated over all τ.
is the conditional cumulative distribution function of ${Y}_{t}$ defined in Eq. (13).
3 Parameter Estimation Algorithms
The unknown parameters in Eq. (3) are α, β, and γ, while we assume Ω known. The reason why the amplitude, γ, is often unknown while the frequency, Ω, is known is that one can usually observe the sinusoidal input and thus its frequency. Further, the encoding of the input into neuronal firing patterns often involves phase locking to the sinusoidal component. However, the actual forcing amplitude at the level of the neuron is usually modified by various synaptic and other filtering processes, unless the cell receives direct sinusoidal current injection.
Our goal is to estimate the structural parameters (α, β, γ) from a sample of spike time data, $\{{i}_{1},\dots ,{i}_{N}\}$. There are several algorithms for estimating the parameters for the simpler and more common case of $\gamma =0$. One such algorithm relies on the Fortet equation (see [7, 8]), which we extend to the presence of a time-varying current. A more basic approach is to directly solve the Fokker–Planck equation for the probability density of ${X}_{t}$, [19–21], from which one can derive the survival distribution of ${I}_{n}$, and use this to compare against the empirical survival distribution of ${I}_{n}$ obtained from data. An approximate maximum likelihood approach is also possible by numerical differentiation. The relation between Fokker–Planck equations and the first-passage time problem is discussed in most introductory books on stochastic analysis; see, for example, [28]. A recent review of this approach for the simple $\gamma =0$ case in neuronal modeling can be found in [21], wherein the first passage problem is discussed at great lengths in the context of spiking neurons. We will use this in Sect. 2.2. A more elaborate approach using the Fokker–Planck equation to approximate the hitting time distribution is given in [29]. The techniques in [29] avoid the need to compute the Fokker–Planck PDE numerically, instead approximating it with analytically known solutions. This approach might offer significant computational savings, but since this would at most amount to a computational speed-up of our algorithm, we have left this unexplored for now.
The immediate problem in generalizing the aforementioned approaches to the case of $\gamma \ne 0$ is that the ${I}_{n}$’s are no longer identically distributed since the phase ${\varphi}_{n-1}$ of the n th interval ${I}_{n}$ depends on ${t}_{n-1}$, the time the previous spike occurred. The ${I}_{n}$’s are also dependent, but conditionally independent given ${\varphi}_{n-1}$. So the trajectories in each interval are parameterized by the value of ${\varphi}_{n-1}$ at the time of the last spike/reset. We overcome this obstacle by splitting the ${I}_{n}$’s in groups, and approximating the ${I}_{n}$’s within groups as coming from identically distributed trajectories in a sense to be specified below. This approximation which solves the challenge of dependent and non-identically distributed ISIs is the primary contribution of this paper.
3.1 ϕ-Binning
Before we can use Eq. (9) or (15), we need to deal with the fact that ϕ is not fixed, but instead each ${I}_{n}$ starts with a distinct ${\varphi}_{n}$. Our approach is to partition the interval $[0,2\pi /\Omega ]$ into M bins, where $M\ll N$, and represent each bin by the midpoint of the bin, ${\varphi}_{m}$. Then we approximate the N observed ${\varphi}_{n}$’s by the closest ${\varphi}_{m}$ and pretend that any observed ${I}_{n}$ was not produced by a trajectory of the form in Eq. (3) with $\varphi ={\varphi}_{n}$, but with $\varphi ={\varphi}_{m}$. Our hope is that for a judicious choice of M, we can balance the error of ${\varphi}_{n}\ne {\varphi}_{m}$ with having enough data points in each bin in order to obtain a useful estimate from Eq. (9) or (15).
3.2 Fokker–Planck Algorithm
The weight ${N}_{m}$ is included so that bins with larger sample sizes have a larger influence on the estimates.
where the derivative has to be approximated by finite differences. We can then again use binning to avoid having to compute the PDE separately for each $({i}_{n},{\varphi}_{n-1})$. Our experience with the MLE approach has been that the quality of the estimates provided are similar to those obtained by minimizing Eq. (17) and that the associated computing times are on the same order. Due to this similarity and in order to keep the paper concise, we include details of the MLE estimates only in the supplementary online material.
3.3 Fortet Algorithm
We divide each inner term by $\omega ({\varphi}_{m};\alpha ,\beta ,\gamma )={sup}_{s>0}|1-{\Phi}_{\alpha ,\beta ,\gamma}^{({\varphi}_{m})}({v}_{\mathrm{th}}(s),s|{v}_{0})|$, following the suggestion in [8]. This scaling ensures that Eq. (15) divided by $\omega (\alpha ,\beta ,\gamma )$ will vary between 0 and 1 for all parameter values thus giving sense to the measure defined by the loss function. Since we can solve in closed form for Φ, we have all we need given an observed spike train of ${i}_{n}$’s. We evaluate the sup by sampling at $K=500$ uniformly spaced points in $(0,{I}_{max}+\u03f5]$ and taking the maximum of the sampled values.
3.4 Initialization of the Algorithms
Its solution given an initial condition $\rho (x,0)=\delta (x)$ will be a Gaussian bell moving to the right with speed U and standard deviation $\sigma =\beta \sqrt{t}$.
for the initializer. We can form these equations separately for each ${\varphi}_{m}$ bin, thus resulting in $M\times 2$ equations for the unknowns α, β, and γ. Since we have more equations than unknowns, we use least-squares estimates in a regression to pick out unique α, β, and γ estimates.
4 Method Comparison on Simulated Data
We will now use our algorithms on spike trains simulated from the four different regimes: the supra-threshold, the critical, the supersinusoidal and the subthreshold. We have used 100 sample spike trains per regime, with $N=100$ as well as $N=1000$ spikes per train. In order to perform the numerical minimization of Eqs. (17) and (18), we have used an implementation of the Nelder–Mead algorithm from the SciPy library [30]. The Nelder–Mead algorithm is a non-linear minimization routine which uses a bounding-polygon method to zero-in on the minimum and thus avoids the need to provide the gradient of the loss function. It is the standard non-gradient minimization algorithm.
Averages and empirical 95 % confidence intervals of the estimates for $N=100$ spikes per train
Parameter | Initializer | Fokker–Planck | Fortet |
---|---|---|---|
Supra-threshold regime | |||
α = 1.40 | 1.43: [1.29,1.56] | 1.34: [1.24,1.43] | 1.41: [1.33,1.49] |
β = 0.30 | 0.17: [0.10,0.24] | 0.29: [0.21,0.39] | 0.29: [0.22,0.36] |
γ = 0.14 | 0.16: [0.02,0.33] | 0.12: [0.02,0.23] | 0.12: [0.01,0.24] |
Supersinusoidal regime | |||
α = 0.10 | 0.92: [0.83,1.01] | 0.28: [0.02,0.59] | 0.24: [−0.22,0.42] |
β = 0.30 | 0.15: [0.10,0.25] | 0.31: [0.14,0.53] | 0.32: [0.14,0.46] |
γ = 1.98 | 1.35: [1.13,1.57] | 1.67: [1.33,2.05] | 1.77: [1.44,2.38] |
Critical regime | |||
α = 0.50 | 0.72: [0.66,0.80] | 0.57: [0.32,0.73] | 0.57: [0.36,0.73] |
β = 0.30 | 0.19: [0.10,0.26] | 0.27: [0.17,0.40] | 0.25: [0.15,0.40] |
γ = 0.71 | 0.57: [0.44,0.73] | 0.55: [0.30,0.83] | 0.62: [0.38,0.93] |
Subthreshold regime | |||
α = 0.40 | 0.62: [0.57,0.67] | 0.63: [0.33,0.84] | 0.58: [0.03,1.00] |
β = 0.30 | 0.17: [0.10,0.29] | 0.20: [0.10,0.37] | 0.19: [0.00,0.41] |
γ = 0.57 | 0.32: [0.00,0.53] | 0.29: [0.00,0.62] | 0.46: [0.00,1.19] |
Averages and empirical 95 % confidence intervals of the estimates for $N=1000$ spikes per train
Parameter | Initializer | Fokker–Planck | Fortet |
---|---|---|---|
Supra-threshold regime | |||
α = 1.40 | 1.44: [1.40,1.50] | 1.36: [1.33,1.40] | 1.40: [1.37,1.42] |
β = 0.30 | 0.25: [0.22,0.28] | 0.29: [0.26,0.32] | 0.30: [0.27,0.32] |
γ = 0.14 | 0.14: [0.10,0.19] | 0.14: [0.10,0.17] | 0.14: [0.10,0.18] |
Supersinusoidal regime | |||
α = 0.10 | 0.90: [0.85,0.92] | 0.11: [0.03,0.29] | 0.10: [0.03,0.16] |
β = 0.30 | 0.18: [0.14,0.23] | 0.30: [0.21,0.34] | 0.31: [0.22,0.34] |
γ = 1.98 | 1.26: [1.16,1.34] | 1.92: [1.49,2.05] | 1.96: [1.86,2.07] |
Critical regime | |||
α = 0.50 | 0.73: [0.70,0.75] | 0.51: [0.43,0.63] | 0.53: [0.45,0.64] |
β = 0.30 | 0.20: [0.17,0.24] | 0.29: [0.24,0.32] | 0.28: [0.19,0.33] |
γ = 0.71 | 0.54: [0.44,0.61] | 0.66: [0.52,0.76] | 0.67: [0.54,0.77] |
Subthreshold regime | |||
α = 0.40 | 0.62: [0.55,0.65] | 0.57: [0.45,0.66] | 0.56: [0.26,0.71] |
β = 0.30 | 0.20: [0.17,0.26] | 0.22: [0.18,0.29] | 0.21: [0.13,0.35] |
γ = 0.57 | 0.36: [0.18,0.44] | 0.36: [0.25,0.50] | 0.43: [0.28,0.72] |
Average times ± standard deviations in seconds for the algorithm in various regimes. Left: $N=100$ spikes; right: $N=1000$ spikes
Regime | Fortet | Fokker–Planck |
---|---|---|
Subthreshold | 1.29 ± 0.72 | 0.52 ± 0.21 |
Supra-threshold | 0.83 ± 0.28 | 0.18 ± 0.20 |
Critical | 0.94 ± 0.42 | 0.36 ± 0.16 |
Supersinusoidal | 1.36 ± 0.46 | 0.43 ± 0.17 |
Subthreshold | 9.68 ± 4.98 | 1.69 ± 0.91 |
Supra-threshold | 3.90 ± 1.05 | 0.21 ± 0.06 |
Critical | 10.03 ± 2.88 | 1.28 ± 0.41 |
Supersinusoidal | 10.13 ± 2.24 | 1.06 ± 0.33 |
5 The Effect of Ω
So far, we have held Ω constant and equal to 1. We now investigate the effect of varying Ω on the quality of estimates. To narrow the scope, we focus on increasing Ω while keeping the parameters in the critical regime such that $\alpha +\gamma /\sqrt{1+{\Omega}^{2}}=1$ and $\alpha =0.5$. This amounts to increasing γ with Ω. We do the estimations for four values of $\Omega =[1,5,10,20]$. Similarly to the previous section, we use 100 sample spike trains per parameter set, with each spike train consisting of $N=1000$ ISIs.
Averages and empirical 95 % confidence intervals of estimates for $N=1000$ spikes per train in the critical regime for varying Ω across $[1,5,10,20]$. Note that the upper subtable corresponds to the third subtable in Table 3; numbers differ slightly due to statistical fluctuations in the simulations
Parameter | Initializer | Fokker–Planck | Fortet |
---|---|---|---|
Ω = 1 | |||
α = 0.50 | 0.73: [0.69,0.75] | 0.52: [0.45,0.61] | 0.52: [0.44,0.62] |
β = 0.30 | 0.20: [0.17,0.25] | 0.29: [0.24,0.33] | 0.29: [0.22,0.34] |
γ = 0.71 | 0.54: [0.44,0.62] | 0.64: [0.53,0.75] | 0.68: [0.55,0.81] |
Ω = 5 | |||
α = 0.50 | 0.88: [0.76,0.99] | 0.78: [0.61,0.89] | 0.64: [0.39,0.99] |
β = 0.30 | 0.24: [0.17,0.31] | 0.26: [0.20,0.34] | 0.27: [0.12,0.34] |
γ = 2.55 | 0.85: [0.00,1.65] | 0.92: [0.00,1.68] | 1.86: [0.00,3.10] |
Ω = 10 | |||
α = 0.50 | 0.90: [0.78,0.99] | 0.71: [0.52,0.88] | 0.58: [0.37,0.86] |
β = 0.30 | 0.25: [0.18,0.33] | 0.26: [0.20,0.35] | 0.28: [0.23,0.32] |
γ = 5.02 | 2.82: [0.92,4.38] | 2.72: [0.95,3.88] | 4.32: [1.20,6.49] |
Ω = 20 | |||
α = 0.50 | 0.93: [0.76,1.02] | 0.75: [0.50,0.92] | 0.62: [0.31,0.97] |
β = 0.30 | 0.27: [0.20,0.33] | 0.29: [0.20,0.43] | 0.29: [0.25,0.33] |
γ = 10.01 | 5.35: [0.00,12.29] | 3.98: [0.00,6.83] | 7.48: [0.00,13.96] |
6 Discussion and Outlook
We have shown two methods to estimate parameters in Eq. (2) from ISI data. Our methods are based on binning the spikes into bins with representative phase shifts. We have devised a constructive procedure to automatically initialize the methods from the data.
Our computational results suggest that for low frequencies the Fortet algorithm is superior for large sample sizes, especially in the supersinusoidal regime, while the Fokker–Planck algorithm has a comparable accuracy and a lower variance for small sample sizes. Both algorithms find sensible estimates most of the time, although they seem less effective in the subthreshold regime. Their performance can be partially attributed to the ability of the initializer algorithm to supply good guesses for starting the optimization iterations.
The Fokker–Planck equation allows for approximate maximum likelihood estimation. We chose an alternative loss function, though, because it marginally appeared more robust, possibly because a numerical derivation step is avoided. This is further investigated by simulations in the supplementary online material. The simulations suggest that the distribution of the maximum likelihood estimates in the supersinusoidal regime appears bimodal, which is not the case for the alternative loss function, Eq. (17).
We have also made a preliminary exploration of the effect of Ω on the quality of the estimates. Our results show that an increase in Ω makes the parameters α and γ more difficult to estimate accurately and at high Ω, γ is underestimated, while α is over-estimated. We find that in this scenario, the Fortet algorithm does a markedly more accurate job then the Fokker–Planck algorithm.
We have assumed the time-constant τ of the leak term to be known. In most experiments that is not realistic, and it would be preferable to estimate τ alongside the other parameters. However, it is difficult to estimate [32]. When we tried to estimate it together with the other parameters, we usually obtained results which were not accurate. The obtained estimates resulted in ISIs that very well matched the data, no worse than the ISIs obtained from the true parameters. This leads us to believe that the simultaneous estimation of τ along with α, β, γ using only ISI data suffers from identifiability problems. In [5], they were able to estimate τ in the simpler nonsinusoidally-driven model, but concluded that adding τ as an unknown dramatically reduced the accuracy in the estimation of the other unknown parameters. The reason is that if τ is also estimated from a single dataset alongside the other parameters, then a reasonable fit can be found to the data for various combinations of α, β, γ, and τ, but the so-obtained parameter values can be far from the true values.
Our model is relatively simple and ignores neurophysiological realism, such as the fact that the spiking threshold is likely nonconstant, with a time-dependent functional form that would involve further unknown parameters. A recent paper attempting the parameter estimation in such a model, but without sinusoidal forcing, is [20]. Furthermore, intracellular recordings suggest that a hard threshold is a rough approximation and an exponential voltage-dependent spiking intensity is more realistic [33].
While our work has used a very specific form of the periodic forcing term, namely $\gamma sin(\Omega t)$, it is clear how to apply the approach to an arbitrary periodic function. This can be done as long as one knows where in the period of oscillation a spike has occurred. If that is the case, then the binning procedure can be applied and the estimation methods proposed can be attempted.
Declarations
Acknowledgements
SD is supported by the Danish Council for Independent Research|Natural Sciences. AL acknowledges support from NSERC Canada. AI is supported by a PGS-D scholarship from NSERC Canada. The work is part of the Dynamical Systems Interdisciplinary Network, University of Copenhagen.
Authors’ Affiliations
References
- Brillinger DR: Maximum likelihood analysis of spike trains of interacting nerve cells. Biol Cybern 1988, 59: 189–200. 10.1007/BF00318010MATHView ArticleGoogle Scholar
- Inoue J, Sato S, Ricciardi L: On the parameter estimation for diffusion models of single neuron’s activities. Biol Cybern 1995, 73: 209–221. 10.1007/BF00201423MATHView ArticleGoogle Scholar
- Ditlevsen S, Lánský P: Estimation of the input parameters in the Ornstein–Uhlenbeck neuronal model. Phys Rev E 2005., 71: Article ID 011907 Article ID 011907Google Scholar
- Ditlevsen S, Lánský P: Estimation of the input parameters in the Feller neuronal model. Phys Rev E 2006., 73: Article ID 061910 Article ID 061910Google Scholar
- Mullowney P, Iyengar S: Parameter estimation for a leaky integrate-and-fire neuronal model from ISI data. J Comput Neurosci 2008, 24(2):179–194. 10.1007/s10827-007-0047-5MathSciNetView ArticleGoogle Scholar
- Zhang X, You G, Chen T, Feng J: Maximum likelihood decoding of neuronal inputs from an interspike interval distribution. Neural Comput 2009, 21(11):3079–3105. 10.1162/neco.2009.06-08-807MATHMathSciNetView ArticleGoogle Scholar
- Ditlevsen S, Ditlevsen O: Parameter estimation from observations of first-passage times of the Ornstein–Uhlenbeck process and the Feller process. Probab Eng Mech 2008, 23: 170–179. 10.1016/j.probengmech.2007.12.024View ArticleGoogle Scholar
- Ditlevsen S, Lánský P: Parameters of stochastic diffusion processes estimated from observations of first-hitting times: application to the leaky integrate-and-fire neuronal model. Phys Rev E 2007., 76: Article ID 041906 Article ID 041906Google Scholar
- Ditlevsen S, Lánský P: Comparison of statistical methods for estimation of the input parameters in the Ornstein–Uhlenbeck neuronal model from first-passage times data. American Institute of Physics Proceedings Series CP1028. In Collective Dynamics: Topics on Competition and Cooperation in the Biosciences. Edited by: Ricciardi LM, Buonocore A, Pirozzi E. Am Inst of Phys, Melville; 2008:171–185.Google Scholar
- Lánský P, Ditlevsen S: A review of the methods for signal estimation in stochastic diffusion leaky integrate-and-fire neuronal models. Biol Cybern 2008, 99(4–5):253–262. 10.1007/s00422-008-0237-xMATHView ArticleGoogle Scholar
- Braun HA, Wissing H, Schafer K, Hirsch MC: Oscillation and noise determine signal-transduction in shark multimodal sensory cells. Nature 1994, 367(6460):270–273. 10.1038/367270a0View ArticleGoogle Scholar
- Chacron MJ, Longtin A, St-Hilaire M, Maler L: Suprathreshold stochastic firing dynamics with memory in P-type electroreceptors. Phys Rev Lett 2000, 85(7):1576–1579. 10.1103/PhysRevLett.85.1576View ArticleGoogle Scholar
- Bulsara AR, Elston TC, Doering CR, Lowen SB, Lindenberg K: Cooperative behavior in periodically driven noisy integrate-fire models of neuronal dynamics. Phys Rev E 1996, 53(4):3958–3969. 10.1103/PhysRevE.53.3958View ArticleGoogle Scholar
- Burkitt AN: A review of the integrate-and-fire neuron model: II. Inhomogeneous synaptic input and network properties. Biol Cybern 2006, 95(2):97–112. 10.1007/s00422-006-0082-8MATHMathSciNetView ArticleGoogle Scholar
- Lánský P: Sources of periodical force in noisy integrate-and-fire models of neuronal dynamics. Phys Rev E 1997, 55(2):2040–2043. 10.1103/PhysRevE.55.2040View ArticleGoogle Scholar
- Longtin A, Bulsara A, Pierson D, Moss F: Bistability and the dynamics of periodically forced sensory neurons. Biol Cybern 1994, 70(6):569–578. 10.1007/BF00198810MATHView ArticleGoogle Scholar
- Sacerdote L, Giraudo MT: Stochastic integrate and fire models: a review on mathematical methods and their applications. Lecture Notes in Mathematics 2058. In Stochastic Biomathematical Models with Applications to Neuronal Modeling. Edited by: Bachar M, Batzel J, Ditlevsen S. Springer, Berlin; 2013.Google Scholar
- Shimokawa T, Pakdaman K, Takahata T, Tanabe S, Sato S: A first-passage-time analysis of the periodically forced noisy leaky integrate-and-fire model. Biol Cybern 2000, 83(4):327–340. 10.1007/s004220000156MATHView ArticleGoogle Scholar
- Paninski L, Pillow J, Simoncelli E: Maximum likelihood estimation of a stochastic integrate-and-fire neural encoding model. Neural Comput 2004, 16(12):2533–2561. 10.1162/0899766042321797MATHView ArticleGoogle Scholar
- Dong Y, Mihalas S, Russell A, Etienne-Cummings R, Niebur E: Estimating parameters of generalized integrate-and-fire neurons from the maximum likelihood of spike trains. Neural Comput 2011, 23(11):2833–2867. 10.1162/NECO_a_00196MATHView ArticleGoogle Scholar
- Sirovich L, Knight B: Spiking neurons and the first passage problem. Neural Comput 2011, 23(7):1675–1703. 10.1162/NECO_a_00139MATHMathSciNetView ArticleGoogle Scholar
- Longtin A: Mechanisms of stochastic phase locking. Chaos, Interdiscip J Nonlinear Sci 1995, 1(5):209–215.View ArticleGoogle Scholar
- Alili L, Patie P, Pedersen J: Representations of the first hitting time density of an Ornstein–Uhlenbeck process. Stoch Models 2005, 21: 967–980. 10.1080/15326340500294702MATHMathSciNetView ArticleGoogle Scholar
- Hurn A, Jeisman J, Lindsay K: ML estimation of the parameters of SDE’s by numerical solution of the Fokker–Planck equation. MODSIM 2005: International Congress on Modelling and Simulation: Advances and Applications for Management and Decision Making 2005, 849–855.Google Scholar
- Karniadakis GE, Kirby RM: Parallel Scientific Computing in C++ and MPI. Cambridge University Press, Cambridge; 2003.MATHView ArticleGoogle Scholar
- Fortet R: Les fonctions aléatoires du type de Markoff associées à certaines équations linéaires aux dérivées partielles du type parabolique. J Math Pures Appl (9) 1943, 22: 177–243.MATHMathSciNetGoogle Scholar
- Shimokawa T, Pakdaman K, Sato S: Time-scale matching in the response of a leaky integrate-and-fire neuron model to periodic stimulus with additive noise. Phys Rev E 1999, 59(3):3427–3443. 10.1103/PhysRevE.59.3427View ArticleGoogle Scholar
- Jacobs K: Stochastic Processes for Physicists: Understanding Noisy Systems. Cambridge University Press, Cambridge; 2010.View ArticleGoogle Scholar
- Lo C, Chung TK: First passage time problem for the Ornstein–Uhlenbeck neuronal model. 13th International Conference on Neural Information Processing 2006, 324–333.View ArticleGoogle Scholar
- Jones E, Oliphant T, Peterson P, et al: SciPy: open source scientific tools for Python; 2001.Google Scholar
- Galasi M: GNU Scientific Library Reference Manual. 3rd edition. Network Theory Ltd, Godalming; 2009.Google Scholar
- Ditlevsen S, Lánský P: Only through perturbation can relaxation times be estimated. Phys Rev E 2012., 86(5): Article ID 050102 Article ID 050102Google Scholar
- Jahn P, Berg RW, Hounsgaard J, Ditlevsen S: Motoneuron membrane potentials follow a time inhomogeneous jump diffusion process. J Comput Neurosci 2011, 31: 563–579. 10.1007/s10827-011-0326-zMathSciNetView ArticleGoogle Scholar
Copyright
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.