 Review
 Open Access
 Published:
Data Assimilation Methods for Neuronal State and Parameter Estimation
The Journal of Mathematical Neuroscience volume 8, Article number: 11 (2018)
Abstract
This tutorial illustrates the use of data assimilation algorithms to estimate unobserved variables and unknown parameters of conductancebased neuronal models. Modern data assimilation (DA) techniques are widely used in climate science and weather prediction, but have only recently begun to be applied in neuroscience. The two main classes of DA techniques are sequential methods and variational methods. We provide computer code implementing basic versions of a method from each class, the Unscented Kalman Filter and 4DVar, and demonstrate how to use these algorithms to infer several parameters of the Morris–Lecar model from a single voltage trace. Depending on parameters, the Morris–Lecar model exhibits qualitatively different types of neuronal excitability due to changes in the underlying bifurcation structure. We show that when presented with voltage traces from each of the various excitability regimes, the DA methods can identify parameter sets that produce the correct bifurcation structure even with initial parameter guesses that correspond to a different excitability regime. This demonstrates the ability of DA techniques to perform nonlinear state and parameter estimation and introduces the geometric structure of inferred models as a novel qualitative measure of estimation success. We conclude by discussing extensions of these DA algorithms that have appeared in the neuroscience literature.
Introduction
The Parameter Estimation Problem
The goal of conductancebased modeling is to be able to reproduce, explain, and predict the electrical behavior of a neuron or networks of neurons. Conductancebased modeling of neuronal excitability began in the 1950s with the Hodgkin–Huxley model of action potential generation in the squid giant axon [1]. This modeling framework uses an equivalent circuit representation for the movement of ions across the cell membrane:
where V is membrane voltage, C is cell capacitance, \(I_{ \textrm{ion}}\) are ionic currents, and \(I_{\textrm{app}}\) is an external current applied by the experimentalist. The ionic currents arise from channels in the membrane that are voltage or calciumgated and selective for particular ions, such sodium (Na^{+}) and potassium (K^{+}). For example, consider the classical Hodgkin–Huxley currents:
The maximal conductance \(g_{\textrm{ion}}\) is a parameter that represents the density of channels in the membrane. The term \((VE_{\textrm{ion}})\) is the driving force, where the equilibrium potential \(E_{\textrm{ion}}\) is the voltage at which the concentration of the ion inside and outside of the cell is at steady state. The gating variable m is the probability that one of three identical subunits of the sodium channel is “open”, and the gating variable h is the probability that a fourth subunit is “inactivated”. Similarly, the gating variable n is the probability that one of four identical subunits of the potassium channel is open. For current to flow through the channel, all subunits must be open and not inactivated. The rate at which subunits open, close, inactivate, and deinactivate depends on the voltage. The dynamics of the gating variables are given by
where \(\alpha_{x}(V)\) and \(\beta_{x}(V)\) are nonlinear functions of voltage with several parameters.
The parameters of conductancebased models are typically fit to voltageclamp recordings. In these experiments, individual ionic currents are isolated using pharmacological blockers and one measures current traces in response to voltage pulses. However, many electrophysiological datasets consist of currentclamp rather than voltageclamp recordings. In currentclamp, one records a voltage trace (e.g., a series of action potentials) in response to injected current. Fitting a conductancebased model to currentclamp data is challenging because the individual ionic currents have not been measured directly. In terms of the Hodgkin–Huxley model, only one state variable (V) has been observed, and the other three state variables (m, h, and n) are unobserved. Conductancebased models of neurons often contain several ionic currents and, therefore, more unobserved gating variables and more unknown or poorly known parameters. For example, a model of HVC neurons in the zebra finch has 9 ionic currents, 12 state variables, and 72 parameters [2]. An additional difficulty in attempting to fit a model to a voltage trace is that if one performs a leastsquares minimization between the data and model output, then small differences in the timing of action potentials in the data and the model can result in large error [3]. Data assimilation methods have the potential to overcome these challenges by performing state estimation (of both observed and unobserved states) and parameter estimation simultaneously.
Data Assimilation
Data assimilation can broadly be considered to be the optimal integration of observations from a system to improve estimates of a model output describing that system. Data assimilation (DA) is used across the geosciences, e.g., in studying land hydrology and ocean currents, as well as studies of climates of other planets [4–6]. An application of DA familiar to the general public is its use in numerical weather prediction [7]. In the earth sciences, the models are typically highdimensional partial differential equations (PDEs) that incorporate dynamics of the many relevant governing processes, and the state system is a discretization of those PDEs across the spatial domain. These models are nonlinear and chaotic, with interactions of system components across temporal and spatial scales. The observations are sparse in time, contaminated by noise, and only partial with respect to the full statespace.
In neuroscience, models can also be highly nonlinear and potentially chaotic. When dealing with network dynamics or wave propagation, the statespace can be quite large, and there are certainly components of the system for which one would not have time course measurements [8]. As mentioned above, if one has a biophysical model of a single neuron and measurements from a currentclamp protocol, the only quantity in the model that is actually measured is the membrane voltage. The question then becomes: how does one obtain estimates of the full system state?
To begin, we assume we have a model to represent the system of interest and a way to relate observations we have of that system to the components of the model. Additionally, we allow, and naturally expect, there to be errors present in the model and measurements. To start, let us consider first a general model with linear dynamics and a set of discrete observations which depend linearly on the system components:
In this statespace representation, \(x_{k}\) is interpreted as the state of the system at some time \(t_{k}\), and \(y_{k}\) are our observations. For application in neuroscience, we can take \(M \ll L\) as few state variables of the system are readily observed. F is our model which maps states \(x_{k}\) between time points \(t_{k}\) and \(t_{k+1}\). H is our observation operator which describes how we connect our observations \(y_{k+1}\) to our statespace at \(t_{k+1}\). The random variables \(\omega_{k+1}\) and \(\eta_{k+1}\) represent model error and measurement error, respectively. A simplifying assumption is that our measurements are diluted by Gaussian white noise, and that the error in the model can be approximated by Gaussian white noise as well. Then \(\omega_{k} \sim\mathcal{N}(0,Q_{k})\) and \(\eta_{k}\sim\mathcal{N}(0,R_{k})\), where \(Q_{k}\) is our model error covariance matrix and \(R_{k}\) is our measurement error covariance matrix. We will assume these distributions for the error terms for the remainder of the paper.
We now have defined a stochastic dynamical system where we have characterized the evolution of our states and observations therein based upon assumed error statistics. The goal is now to utilize these transitions to construct methods to best estimate the state x over time. To approach this goal, it may be simpler to consider the evaluation of background knowledge of the system compared to what we actually observe from a measuring device. Consider the following cost function [9]:
where \(\lVert z \rVert^{2}_{A} =z^{T}A^{1}z\). \(P^{b}\) acts to give weight to certain background components \(x^{b}\), and R acts in the same manner to the measurement terms. The model or background term acts to regularize the cost function. Specifically, trying to minimize \(\frac{1}{2} \lVert yHx \rVert^{2}_{R}\) is underdetermined with respect to the observations unless we can observe the full system, and the model term aims to inform the problem of the unobserved components. We are minimizing over state components x. In this way, we balance the influence of what we think we know about the system, such as from a model, compared to what we can actually observe. The cost function is minimized from
This can be restructured as
where
The optimal Kalman gain matrix K acts as a weighting of the confidence of our observations to the confidence of our background information given by the model. If the background uncertainty is relatively high or the measurement uncertainty is relatively low, K is larger, which more heavily weights the innovation \(yHx^{b}\).
The solution of (7) can be interpreted as the solution of a single time step in our statespace problem (5)–(6). In the DA literature, minimizing this cost function independent of time is referred to as 3DVar. However, practically we are interested in problems resembling the following:
where formally the background component \(x^{b}\) has now been replaced with our model. Now we are concerned with minimizing over an observation window with \(N+1\) time points. Variational methods, specifically “weak 4DVar”, seek minima of (11) either by formulation of an adjoint problem [10], or directly from numerical optimization techniques.
Alternatively, sequential data assimilation approaches, specifically filters, aim to use information from previous time points \(t_{0},t_{1},\ldots ,t_{k}\), and observations at the current time \(t_{k+1}\), to optimally estimate the state at \(t_{k+1}\). The classical Kalman filter utilizes the form of (10), which minimizes the trace of the posterior covariance matrix of the system at step \(k+1\), \(P_{k+1}^{a}\), to update the state estimate and system uncertainty.
The Kalman filtering algorithm takes the following form. Our analysis estimate, \(\hat{x}_{k}^{a}\) from the previous iteration, is mapped through the linear model operator F to obtain our forecast estimate \(\hat{x}^{f}_{k+1}\):
The observation operator H is applied to the forecast estimate to generate the measurement estimate \(\hat{y}^{f}_{k+1}\):
The forecast estimate covariance \(P^{f}_{k+1}\) is generated through calculating the covariance from the model and adding it with the model error covariance \(Q_{k}\):
Similarly, we can construct the measurement covariance estimate by calculating the covariance from our observation equation and adding it to the measurement error covariance \(R_{k}\):
The Kalman gain is defined analogously to (10):
The covariance and the mean estimate of the system are updated through a weighted sum with the Kalman gain:
These equations can be interpreted as a predictor–corrector method, where the predictions of the state estimates are \(\hat{x}^{f}_{k+1}\) with corresponding uncertainties \(P_{k+1}^{f}\) in the forecast. The correction, or analysis, step linearly interpolates the forecast predictions with observational readings.
In this paper we only consider filters, however smoothers are another form of sequential DA that also use observational data from future times \(t_{k+2},\ldots, t_{k+l}\) to estimate the state at \(t_{k+1}\).
Nonlinear Data Assimilation Methods
Nonlinear Filtering
For nonlinear models, the Kalman equations need to be adapted to permit nonlinear mappings in the forward operator and the observation operator:
Our observation operator for voltage data remains linear: \(h(x) = Hx=[\mathbf{e_{1}} 0 \dots0] x\), where ej is the jth elementary basis vector, is a projection onto the voltage component of our system. Note that \(h(x)\) is an operator, not to be confused with the inactivation gate in (2). Our nonlinear model update, \(f(x)\) in (19), is taken as the forward integration of the dynamical equations between observation times.
Multiple platforms for adapting the Kalman equations exist. The most straightforward approach is the extended Kalman filter (EKF) which uses local linearizations of the nonlinear operators in (19)–(20) and plugs these into the standard Kalman equations. By doing so, one preserves Gaussianity of the statespace. Underlying the data assimilation framework is the goal of understanding the distribution, or statistics of the distribution, of the states of the system given the observations:
The Gaussianity of the statespace declares the posterior conditional distribution \(p(x  y)\) to be a normal distribution by the product of Gaussians being Gaussian, and the statistics of this distribution lead to the Kalman update equations [10]. However, the EKF is really only suitable when the dynamics are nearly linear between observations and can result in divergence of the estimates [11].
Rather than trying to linearize the transformation to preserve Gaussianity, where this distributional assumption is not going to be valid for practical problems anyway, an alternative approach is to preserve the nonlinear transformation and try to estimate the first two moments of transformed state [11]. The Unscented Kalman Filter (UKF) approximates the first two statistics of \(p(x_{k}  y_{0}\ldots y_{k} )\) by calculating sample means and variances, which bypasses the need for Gaussian integral products. The UKF uses an ensemble of deterministically selected points in the statespace whose collective mean and covariance are that of the state estimate and its associated covariance at some time. The forward operator \(f(x)\) is applied to each of these sigma points, and the mean and covariance of the transformed points can then be computed to estimate the nonlinearly transformed mean and covariance. Figure 1 depicts this “unscented” transformation. The sigma points precisely estimate the true statistics both initially (Fig. 1(A)) and after nonlinear mapping (Fig. 1(B)).
In the UKF framework, as with all DA techniques, one is attempting to estimate the states of the system. The standard set of states in conductancebased models includes the voltage, the gating variables, and any intracellular ion concentrations not taken to be stationary. To incorporate parameter estimation, parameters θ to be estimated are promoted to states whose evolution is governed by the model error random variable:
This is referred to as an “artificial noise evolution model”, as the random disturbances driving deviations in model parameters over time rob them of their timeinvariant definition [12, 13]. We found this choice to be appropriate for convergence and as a tuning mechanism. An alternative is to zero out the entries of \(Q_{k}\) corresponding to the parameters in what is called a “persistence model” where \(\boldsymbol {\theta}_{k+1}= \boldsymbol{\theta}_{k}\) [14]. However, changes in parameters can still occur during the analysis stage.
We declare our augmented state to be comprised of the states in the dynamical system as well as parameters θ of interest:
where q represents the additional states of the system besides the voltage. The filter requires an initial guess of the state \(\hat{x}_{0}\) and covariance \(P_{xx}\). An implementation of this algorithm is provided as Supplementary Material with the parent function UKFML.m and one time step of the algorithm computed in UKF_Step.m.
An ensemble of σ points are formed and their position and weights are determined by λ, which can be chosen to try to match higher moments of the system distribution [11]. Practically, this algorithmic parameter can be chosen to spread the ensemble for \(\lambda>0\), shrink the ensemble for \(N< \lambda< 0\), or to have the mean point completely removed from the ensemble by setting it to zero. The ensemble is formed on lines 8082 of UKF_Step.m. The individual weights can be negative, but their cumulative sum is 1.
We form our background estimate \(\hat{x}_{k+1}^{b}\) by applying our map \(f(x)\) to each of the ensemble members
and then computing the resulting mean:
We then propagate the transformed sigma points through the observation operator
and compute our predicted observation \(\hat{y}_{k+1}^{b}\) from the mapped ensemble:
We compute the background covariance estimate by calculating the variance of the mapped ensemble and adding the process noise \(Q_{k}\):
and do the same for the predicted measurement covariance with the addition of \(R_{k}\):
The Kalman gain is computed by matrix multiplication of the crosscovariance:
with the predicted measurement covariance:
When only observing voltage, this step is merely scalar multiplication of a vector. The gain is used in the analysis, or update step, to linearly interpolate our background statistics with measurement corrections. The update step for the covariance is
and the mean is updated to interpolate the background estimate with the deviations of the estimated measurement term with the observed data \(y_{k+1}\):
The analysis step is performed on line 124 of UKF_Step.m. Some implementations also include a redistribution of the sigma points about the forecast estimate using the background covariance prior to computing the crosscovariance \(P_{xy}\) or the predicted measurement covariance \(P_{yy}\) [15]. So, after (29), we redefine \(\tilde{X_{j}}\), \(\tilde{Y}_{j}\) in (25) as follows:
The above is shown in lines 98–117 in UKF_Step. A particularly critical part of using a filter, or any DA method, is choosing the process covariance matrix \(Q_{k}\) and the measurement covariance matrix \(R_{k}\). The measurement noise may be intuitively based upon knowledge of one’s measuring device, but the model error is practically impossible to know a priori. Work has been done to use previous innovations to simultaneously estimate Q and R during the course of the estimation cycle [16], but this becomes a challenge for systems with low observability (such as is the case when only observing voltage). Rather than estimating the states and parameters simultaneously as with an augmented statespace, one can try to estimate the states and parameters separately. For example, [17] used a shooting method to estimate parameters and the UKF to estimate the states. This study also provided a systematic way to estimate an optimal covariance inflation \(Q_{k}\). For highdimensional systems where computational efficiency is a concern, an implementation which efficiently propagates the square root of the state covariance has been developed [18].
Figure 2 depicts how the algorithm operates. Between observation times, the previous analysis (or best estimate) point is propagated through the model to come up with the predicted model estimate. The Kalman update step interpolates this point with observations weighted by the Kalman gain.
Variational Methods
In continuous time, variational methods aim to find minimizers of functionals which represent approximations to the probability distribution of a system conditioned on some observations. As our data is available only in discrete measurements, it is practical to work with a discrete form similar to (7) for nonlinear systems:
We assume that the states follow the statespace description in (19)–(20) with \(\omega_{k}\sim \mathcal{N}(0,Q)\) and \(\eta_{k}\sim\mathcal{N}(0,R)\), where Q is our model error covariance matrix and R is our measurement error covariance matrix. As an approximation, we impose Q, R to be diagonal matrices, indicating that there is assumed to be no correlation between errors in other states. Namely, Q, contains only the assumed model error variance for each statespace component, and R is just the measurement error variance of the voltage observations. These assumptions simplify the cost function to the following:
where \(V_{k} =x_{1,k}\). For the currentclamp data problem in neuroscience, one seeks to minimize equation (36) in what is called the “weak 4DVar” approach. An example implementation of weak 4DVar is provided in w4DvarML.m in the Supplementary Material. An example of the cost function with which to minimize over is given in the child function w4dvarobjfun.m. Each of the \(x_{k}\) is mapped by \(f(x)\) on line 108. Alternatively, “strong 4DVar” forces the resulting estimates to be consistent with the model \(f(x)\). This can be considered the result of taking \(Q \rightarrow\mathbf{0}\), which yields the nonlinearly constrained problem
such that
The rest of this paper will be focused on the weak case (36), where we can define the argument of the optimization as follows:
resulting in an \((N+1)L + D\)dimensional estimation problem. An important aspect of the scalability of this problem is that the Hessian matrix
is sparse. Namely, each state at each discrete time has dependencies based upon the model equations and the chosen numerical integration scheme. At the heart of many gradientbased optimization techniques lies a linear system, involving the Hessian and the gradient \(\nabla C(\mathbf{x}_{n})\) of the objective function, that is used to solve for the next candidate point. Specifically, Newton’s method for optimization is
Therefore, if \((N+1)L + D\) is large, then providing the sparsity pattern is advantageous when numerical derivative approximations, or functional representations of them, are being used to perform minimization with a derivativebased method. One can calculate these derivatives by hand, symbolic differentiation, or automatic differentiation.
A feature of the most common derivativebased methods is assured convergence to local minima. However, our problem is nonconvex due to the model term, which leads to the development of multiple local minima in the optimization surface as depicted in Fig. 3. For the results in this tutorial, we will only utilize local optimization tools, but see Sect. 5 for a brief discussion of some global optimization methods with stochastic search strategies.
Application to Spiking Regimes of the Morris–Lecar Model
Twin Experiments
Data assimilation is a framework for the incorporation of system observations into an estimation problem in a systematic fashion. Unfortunately, the methods themselves do not provide a great deal of insight into the tractability of unobserved system components of specific models. There may be a certain level of redundancy in the model equations and degeneracy in the parameter space leading to multiple potential solutions [19]. Also, it may be the case that certain parameters are nonidentifiable if, for instance, a parameter can be completely scaled out [20]. Some further work on identifiability is ongoing [21, 22].
Before applying a method to data from a real biological experiment, it is important to test it against simulated data where the ground truth is known. In these experiments, one creates simulated data from a model and then tries to recover the true states and parameters of that model from the simulated data alone.
Recovery of Bifurcation Structure
In conductancebased models, as well as in real neurons, slight changes in a parameter value can lead to drastically different model output or neuronal behavior. Sudden changes in the topological structure of a dynamical system upon smooth variation of a parameter are called bifurcations. Different types of bifurcations lead to different neuronal properties, such as the presence of bistability and subthreshold oscillations [23]. Thus, it is important for a neuronal model to accurately capture the bifurcation dynamics of the cell being modeled [24]. In this paper, we ask whether or not the models estimated through data assimilation match the bifurcation structure of the model that generated the data. This provides a qualitative measure of success or failure for the estimation algorithm. Since bifurcations are an inherently nonlinear phenomenon, our use of topological structure as an assay emphasizes how nonlinear estimation is a fundamentally distinct problem from estimation in linear systems.
Morris–Lecar Model
The Morris–Lecar model, first used to describe action potential generation in barnacle muscle fibers, has become a canonical model for studying neuronal excitability [25]. The model includes an inward voltagedependent calcium current, an outward voltagedependent potassium current, and a passive leak current. The activation gating variable for the potassium current has dynamics, whereas the calcium current activation gate is assumed to respond instantaneously to changes in voltage. The calcium current is also noninactivating, resulting in a twodimensional model. The model exhibits multiple mechanisms of excitability: for different choices of model parameters, different bifurcations from quiescence to repetitive spiking occur as the applied current is increased [23]. Three different bifurcation regimes—Hopf, saddlenode on an invariant circle (SNIC), and homoclinic—are depicted in Fig. 4 and correspond to the parameter sets in Table 1. For a given applied current in the region where a stable limit cycle (corresponding to repetitive spiking) exists, each regime displays a distinct firing frequency and action potential shape.
The equations for the Morris–Lecar model are as follows:
with
The eight parameters that we will attempt to estimate from data are \(g_{\textrm{L}}\), \(g_{\textrm{K}}\), \(g_{\textrm{{Ca}}}\), ϕ, \(V_{1}\), \(V_{2}\), \(V_{3}\), and \(V_{4}\). We are interested in whether the estimated parameters yield a model with the desired mechanism of excitability. Specifically, we will conduct twin experiments where the observed data is produced by a model with parameters in a certain bifurcation regime, but the data assimilation algorithm is initialized with parameter guesses corresponding to a different bifurcation regime. We then assess whether or not a model with the set of estimated parameters undergoes the same bifurcations as the model that produced the observed data. This approach provides an additional qualitative measure of estimation accuracy, beyond simply comparing the values of the true and estimated parameters.
Results with UKF
The UKF was tested on the Morris–Lecar model in an effort to simultaneously estimate V and n along with the eight parameters in Table 1. Data was generated via a modified Euler scheme at observation points every 0.1 ms, where we take the stepsize Δt as 0.1 as well:
The UKF is a particularly powerful tool when a lot of data is available; the computational complexity in time is effectively the same as the numerical scheme of choice, whereas the additional operations at each time point are \(O((L+D)^{3})\) [26]. \(f(x)\) in (19) is taken to be the Morris–Lecar equations (42)–(43), acting as \(f^{\star}(t_{k},x_{k})\), integrated forward via modified Euler (47), and is given on line 126 of UKFML.m. The function fXaug.m, provided in the Supplementary Material, represents our augmented vector field. Our observational operator H is displayed on line 136 of UKFML.m. To reiterate, the states to be estimated in the Morris–Lecar model are the voltage and the potassium gating variable. The eight additional parameters are promoted to the members of statespace with trivial dynamics resulting in a tendimensional estimation problem.
These examples were run using 20 seconds of data which is 200,001 time points. During this time window, the Hopf, SNIC, and homoclinic models fire 220, 477, and 491 spikes, respectively. Such a computation for a tendimensional model takes only a few minutes on a laptop computer. R can be set to 0 when one believes the observed signal to be completely noiseless, but even then it is commonly left as a small number to try to mitigate the development of singularities in the predicted measurement covariance. We set our observed voltage to be the simulated output using modified Euler with additive white noise at each time point:
where \(\eta\sim\mathcal{N}(0,(\varepsilon \sigma_{\textrm{true}})^{2})\) is a normal random variable whose variance is equal to the square of the standard deviation of the signal scaled by a factor ε, which is kept fixed at 0.01 for these simulations. R is taken as the variance of η. The initial covariance of the system is \(\alpha_{I}I\), where I is the identity matrix and \(\alpha_{I}\) is 0.001. The initial guess for n is taken to be 0. Q is fixed in time as a diagonal matrix with diagonal 10^{−7} \([ \max(V_{\textrm{obs}})\min(V_{\textrm{obs}}), 1,\lvert \boldsymbol{\theta_{0}}\rvert ] \), where \(\boldsymbol{\theta_{0}}\) represents our initial parameter guesses. We set \(\lambda=5\); however, this parameter was not especially influential for the results of these runs, as discussed further below. These initializations are displayed in the body of the parent function UKFML.m.
Figure 5 shows the state estimation results when the observed voltage is from the SNIC regime, but the UKF is initialized with parameter guess corresponding to the Hopf regime. Initially, the state estimate for n and its true, unobserved dynamics have great disparity. As the observations are assimilated over the estimation window, the states and model parameters adjust to produce estimates which better replicate the observed, and unobserved, system components. In this way, information from the observations is transferred to the model. The evolution of the parameter estimates for this case is shown in the first column of Fig. 6, with ϕ, \(V_{3}\), and \(V_{4}\) all converging to close to their true values after 10 seconds of observations. The only difference in parameter values between the SNIC and homoclinic regimes is the value of the parameter ϕ. The second column of Fig. 6 shows that when the observed data is from the homoclinic regime but the initial parameter guesses are from the SNIC regime, the estimates of \(V_{3}\) and \(V_{4}\) remain mostly constant near their original (and correct) values, whereas the estimate of ϕ quickly converges to its new true value. Finally, the third column of Fig. 6 shows that all three parameter estimates evolve to near their true values when the UKF is presented with data from the Hopf regime but initial parameter estimates from the homoclinic regime.
Table 2 shows the parameter estimates at the end of the estimation window for all of the nine possible twin experiments. Promisingly, a common feature of the results is the near recovery of the true value of each of the parameters. However, the estimated parameter values alone do not necessarily tell us about the dynamics of the inferred model. To assess the inferred models, we generate bifurcation diagrams using the estimated parameters and compare them to the bifurcation diagrams for the parameters that produced the observed data. Figure 7 shows that the SNIC and homoclinic bifurcation diagrams were recovered quite exactly. The Hopf structure was consistently recovered, but with shifted regions of spiking and quiescence and minor differences in spike amplitude.
To check the consistency of our estimation, we set 100 initial guesses for n across its dynamical range as samples from \(\mathcal{U}(0,1)\). Figure 8 shows that the state estimates for n across these initializations quickly approached very similar trajectories. We confirmed that after the estimation cycle was over, the parameter estimates for all 100 initializations were essentially identical to the values shown in Table 2. In this paper, we always initialized the UKF with initial parameter values corresponding to the various bifurcation regimes and did not explore the performance for randomly selected initial parameter guesses. For initial parameter guesses that are too far from the true values, it is possible that the filter would converge to incorrect parameter values or fail outright before reaching the end of the estimation window. Additionally, we investigated the choices of certain algorithmic parameters for the UKF, namely λ and \(\alpha_{I}\). Figure 9(A) shows suitable ranges of these parameters, with the color indicating the root mean squared error of the parameters at the end of the cycle compared to their true values. We found this behavior to be preserved across our nine twin experiment scenarios. Notably, this shows that our results in Table 2 were generated using an initial covariance \(\alpha_{I}=0.001\) that was smaller than necessary. By increasing the initial variability, the estimated system can converge to the true dynamics more quickly, as shown for \(\alpha_{I}=0.1\) in Fig. 9(B). The value of λ does not have a large impact on these results, except for when \(\alpha_{I}=1\). Here the filter fails before completing the estimation cycle, except for a few cases where λ is small enough to effectively shrink the ensemble spread and compensate for the large initial covariance. For example, with \(\lambda=9\), we have \(N9=1\) and, therefore, the ensemble spread in (24) is simply \(X_{j}=\hat{x}^{a}_{k}\pm\sqrt{P_{xx}}\). For even larger initial covariances (\(\alpha_{I}>1\)), the filter fails regardless of the value of λ. We noticed that in many of the cases that failed, the parameter estimate for ϕ was becoming negative (which is unrealistic for a rate) or quite large (\(\phi>1\)), and that the state estimate for n was going outside of its biophysical range of 0 to 1. When the gating variable extends outside of its dynamical range it can skew the estimated statistics and the filter may be unable to recover. The standard UKF framework does not provide a natural way of incorporating bounds on parameter estimates, and we do not apply any for the results presented here. However, we did find that we can modify our numerical integration scheme to prevent the filter from failing in many of these cases, as shown in Fig. 9(C). Specifically, if n becomes negative or exceeds 1 after the update step, then artificially setting n to 0 or 1 in the modified Euler method (47) before proceeding can enable the filter to reach the end of the estimation window and yield reasonable parameter estimates.
Results with 4DVar
The following results illustrate the use of weak 4DVar. One can minimize the cost function (36) using a favorite choice of optimization routine. For the following examples, we will consider a local optimizer by using interior point optimization with MATLAB’s builtin solver fmincon. At the heart of the solver is a Newtonstep which uses information about the Hessian, or a conjugate gradient step using gradient information [27–29]. The input we are optimizing over conceptually takes the form of
resulting in an \((N+1)L + D\)dimensional estimation problem where \(L=2\). There are computational limitations with memory storage and the time required to sufficiently solve the optimization problem to a suitable tolerance for reasonable parameter estimates. Therefore, we cannot be cavalier with using as much data with 4DVar as we did with the UKF, as that would result in a \((200\text{,}001)2 + 8 = 400\text{,}010\) dimensional problem. Using Newton’s method (41) on this problem would involve inverting a Hessian matrix of size \((400\text{,}010)^{2}\), which according to a rough calculation would require over 1 TB of RAM. Initialization of the optimization is shown on line 71 of w4DVarML.m.
The estimated parameters are given in Table 3. These results were run using \(N=2001\) time points. To simplify the search space, the parameter estimates were constrained between the bounds listed in Table 4. These ranges were chosen to ensure that the maximal conductances, the rate ϕ, and the activation curve slope \(V_{2}\) all remain positive. We found that running 4DVar with even looser bounds (Table A1) yielded less accurate parameter estimates (Tables A2 and A3). The white noise perturbations for the 4DVar trials were the same as those from the UKF examples. Initial guesses for the states at each time point are required. For these trials, V is initialized as \(V_{\textrm{obs}}\), and n is initialized as the result of integration of its dynamics forced with \(V_{\textrm{obs}}\) using the initial guesses for the parameters, i.e., \(n = \int f_{n}(V_{\textrm{obs}},n; \boldsymbol{\theta_{0})}\). The initial guesses are generated beginning on line 38 of w4DvarML.m. We impose that \(Q^{1}\) in (36) is a diagonal matrix with entries \(\alpha_{Q} [1, 100^{2}]\) to balance the dynamical variance of V and n. The scaling factor \(\alpha_{Q}\) represents the relative weight of the model term compared to the measurement term. Based on preliminary tuning experiments, we set \(\alpha_{Q}=100\) for the results presented.
Figure 10 depicts the states produced by integrating the model with the estimated parameters across different iterations within the interiorpoint optimization. Over iteration cycles, the geometry of spikes as well as the spike time alignments eventually coincide with the noiseless data \(V_{\textrm{true}}\). Figure 11 shows the evolution of the parameters across the entire estimation cycle. For the UKF, the “plateauing” effect of the parameter estimates seen in Fig. 6 indicates confidence that they are conforming to being constant in time. With 4DVar, and in a limiting sense of the UKF, the plateauing effect indicates the parameters are settling into a local minimum of the cost function.
In Fig. 12 we show the bifurcation diagrams of the estimated models from our 4DVar trials. Notice, and shown explicitly in Table 3, when initializing with the true parameters, the correct model parameters are recovered as our optimization routine is confidently within the basin of attraction of the global minimum. In the UKF, comparatively, there is no sense of stopping at a local minimum. Parameter estimates may still fluctuate even when starting from their true values, unless the variances of the state components fall to very low values and the covariance \(Q_{k}\) can be tuned to have a baseline variability in the system. The parameter sets for the SNIC and homoclinic bifurcation regimes only deviate in the ϕ parameter, and so our optimization had great success estimating one from the other. The kinetic parameters (\(V_{3}\) and \(V_{4}\)) for the Hopf regime deviate quite a bit from the SNIC or homoclinic. Still, the recovered bifurcation structures from estimated parameters associated with trials involving HOPF remained consistent with the true structure.
A drawback of the results shown in Table 3 is that for the default tolerances in fmincon, some runs took more than two days to complete on a dedicated core. Figure 11 shows that the optimal solution had essentially been found after 22,000 iterations; however, the optimizer kept running for several thousand more iterations before the convergence tolerances were met. Rather than attempting to speed up these computations by adjusting the algorithmic parameters associated with this solver for this specific problem, we decided to try to exploit the dynamic structure of the model equations using automatic differentiation (AD). AD deconstructs derivatives of the objective function into elementary functions and operations through the chain rule. We used the MATLAB AD tool ADiGator, which performs source transformation via operator overloading and has scripts available for simple integration with various optimization tools, including fmincon [30]. For the same problem scenario and algorithmic parameters, we additionally passed in the generated gradient and Hessian functions to the solver. For this problem, the Hessian structure is shown in Fig. 13. Note that we are using a very simple scheme in the modified Euler method (47) to perform numerical integration between observation points, and the states at \(k+1\) only have dependencies upon those at k and on the parameters. Higher order methods, including implicit methods, can be employed naturally since the system is being estimated simultaneously. A tutorial specific to collocation methods for optimization has been developed [31].
The results are shown in Table A4. Each twin experiment scenario took, at most, a few minutes on a dedicated core. These trials converged to the optimal solution in much fewer iterations than the trials without using the Hessian. Since convergence was achieved within a few dozen iterations, we decided to inspect how the bifurcation structure of the estimated model evolved throughout the process for the case of HOPF to SNIC. Figure 14 shows that by Iteration 10, the objective function value has decreased greatly, and parameters that produce a qualitatively correct bifurcation structure have been found. The optimization continues for another 37 iterations and explores other parts of parameter space that do not yield the correct bifurcation structure before converging very close to the true parameter values.
Again, these results, at best, can reflect only locally optimal solutions of the optimization manifold. The 4DVar framework has been applied to neuroscience using a more systematic approach to finding the global optimum. In [32], a population of initial states x is optimized in parallel with an outer loop that incorporates an annealing algorithm. The annealing parameter relates the weights of the two summations in (36), and the iteration proceeds by increasing the weight given to the model error compared to the measurement error.
We also wished to understand more about the sensitivity of this problem to initial conditions. We initialized the system with the voltage states as those of the observation, the parameters as those of the initializing guess bifurcation regime, and the gating variable \([n_{0}, n_{1},\dots n_{N}]\) to be i.i.d. from \(\mathcal{U}(0,1)\). The results confirm our suspicions that multiple local minima exist. For 100 different initializations of n, for the problem of going from SNIC to HOPF, 63 were found to fall into a deeper minima, yielding better estimates and a smaller objective function value, while 16 fell into a shallower minima, and the rest into three different even shallower minima. While one cannot truly visualize highdimensional manifolds, one can try to visualize a subset of the surface. Figure 3 shows the surface that arises from evaluating the objective function on a linear combination of the two deepest minima and an initial condition \(\mathbf{x}_{0}\), which eventually landed in the shallower of the two minima as points in 4010dimensional space.
Application to Bursting Regimes of the Morris–Lecar Model
Many types of neurons display burst firing, consisting of groups of spikes separated by periods of quiescence. Bursting arises from the interplay of fast currents that generate spiking and slow currents that modulate the spiking activity. The Morris–Lecar model can be modified to exhibit bursting by including a calciumgated potassium (K\(_{ \textrm{Ca}}\)) current that depends on slow intracellular calcium dynamics [33]:
Bursting can be analyzed mathematically by decomposing models into fast and slow subsystems and applying geometric singular perturbation theory. Several different types of bursters have been classified based on the bifurcation structure of the fast subsystem. In squarewave bursting, the active phase of the burst is initiated at a saddlenode bifurcation and terminates at a homoclinic bifurcation. In elliptic bursting, spiking begins at a Hopf bifurcation and terminates at a saddlenode of periodic orbits bifurcation. The voltage traces produced by these two types of bursting are quite distinct, as shown in Fig. 15.
Results with UKF
We conducted a set of twin experiments for the bursting model to address the same question as we did for the spiking model: from a voltage trace alone, can DA methods estimate parameters that yield the appropriate qualitative dynamical behavior? Specifically, we simulated data from the squarewave (elliptic) bursting regime, and then initialized the UKF with parameter guesses corresponding to elliptic (squarewave) bursting (these parameter values are shown in Table 5). As a control experiment, we also ran the UKF with initial parameter guesses corresponding to the same bursting regime as the observed data. The observed voltage trace included additive white noise generated following the same protocol as in previous trials. We used 200,001 time points with observations at every 1 ms. Between observations, the system was integrated forward using substeps of 0.025 ms. For the squarewave burster, this included 215 bursts with 4 spikes per burst, and 225 bursts with 2 spikes for the elliptic burster.
The small parameters ε and μ in the calcium dynamics equation were assumed to be known and were not estimated by the UKF. Thus, for the bursting model, we are estimating one additional state variable (Ca) and one additional parameter (\(g_{\textrm{KCa}}\)) compared to the case for the spiking model. Table 6 shows the UKF parameter estimates after initialization with either the true parameters or the parameters producing the other type of bursting. The results for either case are quite consistent and fairly close to their true values for both types of bursting. Since small changes in parameter values can affect bursting dynamics, we also computed bifurcation diagrams for these estimated parameters and compared them to their true counterparts. Figure 16 shows that in all four cases, the estimated models have the same qualitative bifurcation structure as the models that produced the data. The recovered parameter estimates were insensitive to the initial conditions for n and Ca, with 100 different initializations for these state variables sampled from \(\mathcal{U}(0,1)\) and \(\mathcal {U}(0,5)\), respectively. Note, most predominantly in the top right panel, the location of the bifurcations is relatively sensitive to small deviations in certain parameters, such as \(g_{\textrm{KCa}}\). Estimating \(g_{\textrm{KCa}}\) is challenging due to the algebraic degeneracy of estimating both terms involved in the conductance \(G_{\textrm{KCa}} = g_{\textrm{KCa}}Ca/(Ca+1)\), and the inherent timescale disparity of the Ca dynamics compared to V and n. If one had observations of calcium, or full knowledge of its dynamical equations, this degeneracy would be immediately alleviated. To address difficulties in the estimation of bursting models, an approach that separates the estimation problem into two stages based on timescales—first estimating the slow dynamics with the fast dynamics blocked and then estimating the fast dynamics with the slow parameters held fixed—has been developed [34].
Results with 4DVar
We also investigated the utility of variational techniques to recover the mechanisms of bursting. For these runs, we took our observations to be coarsely sampled at 0.1 ms, and our forward mapping is taken to be one step of modified Euler between observation times, as was the case for our previous 4DVar Morris–Lecar results. We used 10,000 time points, which is one burst for the square wave burster, and one full burst plus another spike for the elliptic burster. We used the LBFGSB method [35], as we found it to perform faster for this problem than fmincon. This method approximates the Broyden–Fletcher–Goldfarb–Shanno (BFGS) quasiNewton algorithm using a limited memory (L) inverse Hessian approximation, with an extension to handle bound constraints (B). It is available for Windows through the OPTI toolbox [36] or through a nonspecific operating system MATLAB MEX wrapper [37]. We supplied the gradient of the objective function, but allowed the solver to define the limitedmemory Hessian approximation for our 30,012dimensional problem. The results are captured in Table 7. We performed the same tests with providing the Hessian; however, there was no significant gain in accuracy or speed. The value for \(g_{\mathrm{KCa}}\) for initializing with the square wave parameters and estimating the elliptical parameters is quite off, which reflects our earlier assessment for the value in observing calcium dynamics. Figure 17 shows that we are still successful in recovering the true bifurcation structure.
Discussion and Conclusions
Data assimilation is a framework by which one can optimally combine measurements and a model of a system. In neuroscience, depending on the neural system of interest, the data we have may unveil only a small subset of the overall activity of the system. For the results presented here, we used simulated data from the Morris–Lecar model with distinct activity based upon different choices for model parameters. We assumed access only to the voltage and the input current, which corresponds to the expected data from a currentclamp recording.
We showed the effectiveness of standard implementations of the Unscented Kalman Filter and weak 4DVar to recover spiking behavior and, in many circumstances, nearexact parameters of interest. We showed that the estimated models undergo the same bifurcations as the model that produced the observed data, even when the initial parameter guesses do not. Additionally, we are also provided with estimates of the states and uncertainties associated with each state and parameter, but for sake of brevity these values were not always displayed. The methods, while not insensitive to noise, have intrinsic weightings of measurement deviations to account for the noise of the observed signal. Results were shown for mild additive noise. We also extended the Morris–Lecar model to exhibit bursting activity and demonstrated the ability to recover these model parameters using the UKF.
The UKF and 4DVar approaches implemented here both attempt to optimally link a dynamic model of a system to observed data from that system, with error statistics assumed to be Gaussian. Furthermore, both approaches try to approximate the mean (and for the UKF also the variance) of the underlying, unassumed system distributions. The UKF is especially adept at estimating states over long time courses, and if the algorithmic parameters such as the model error can be tuned, then the parameters can be estimated simultaneously. Therefore, if one has access to a long series of data, then the UKF (or an Unscented Kalman Smoother, which uses more history of the data for each update step) is a great tool to have at one’s disposal. However, sometimes one only has a small amount of time series data, or the tuning of initial covariance, the spread parameter λ, and the process noise \(Q_{k}\) associated with the augmented state and parameter system becomes too daunting. The 4DVar approach sets the states at each time point and the parameters as optimization variables, transitioning the estimation process from the one which iterates in time to the one which iterates up to a tolerance in a chosen optimization routine. The only tuning parameters are those associated with the chosen optimization routine, and the weights \(Q_{l,l}^{1} , l\in[1 \dots L]\), for the model uncertainty of the state components at each time. There are natural ways to provide parameter bounds in the 4DVar framework, whereas this is not the case for the UKF. However, depending upon the implementation choices and the dimension of the problem (which is extremely large for long time series data), the optimization may take a computing time scale of days to yield reasonable estimates. Fortunately, derivative information can be provided to the optimizer to speed up the 4DVar procedure. Both the UKF and 4DVar can provide estimates of the system uncertainty in addition to estimates of the system mean. The UKF provides mean and variance estimates at each iteration during the analysis step. In 4DVar, one seeks mean estimates by minimization of a cost function. It has been shown that for cost functions of the form (36), the system variance can be interpreted as the inverse of the Hessian evaluated at minima of (36), and scales roughly as Q for large \(Q^{1}\) [32]. The pros and cons of implementing these two DA approaches are summarized in Table 8.
The UKF and 4DVar methodologies welcome the addition of any observables of the system, but currentclamp data may be all that is available. With this experimental data in mind, for a more complex system, the number of variables increases, while the total number of observables will remain at unity. Therefore, it may be useful to assess a priori which parameters are structurally identifiable and the sensitivity of the model to parameters of interest in order to reduce the estimation statespace [38]. Additionally, one should consider what manner of applied current to use to aid in state and parameter estimation. In the results presented above, we used a constant applied current, but work has been done which suggests the use of complex timevarying currents that stimulate as many of the model’s degrees of freedom as possible [39].
The results we presented are based on MATLAB implementations of the derived equations for the UKF and weak 4DVar. Sample code is provided in the Supplementary Material. Additional data assimilation examples in MATLAB can be found in [40]. The UKF has been applied to other spiking neuron models such as the FitzHugh–Nagumo model [41]. A sample of this code can be found in [42], as well as further exploration of the UKF in estimating neural systems. The UKF has been used on real data from pyramidal neurons to track the states and externally applied current [43], the connectivity of cultured neuronal networks sampled by a microelectrode array [44], to assimilate seizure data from hippocampal OLM interneurons [15], and to reconstruct mammalian sleep dynamics [17]. A comparative study of the efficacy of the EKF and UKF on conductancebased models has been conducted [45].
The UKF is a particularly good framework for the state dimensions of a single compartment conductance based model as the size of the ensemble is chosen to be \(2(L+D)+1\). When considering larger state dimensions, as is the case for PDE models, a more general Ensemble Kalman Filter (EnKF) may be appropriate. An introduction to the EnKF can be found in [46, 47]. An adaptive methodology using past innovations to iteratively estimate the model and measurement covariances Q and R has been developed for use with ensemble filters [16]. The Local Ensemble Tranform Kalman Filter (LETKF) [48] has been used to estimate the states associated with cardiac electrical wave dynamics [8]. Rather than estimating the mean and covariance through an ensemble, particle filters aim to fully construct the posterior density of the states conditioned on the observations. A particle filter approach has been applied to infer parameters of a stochastic Morris–Lecar model [49], to assimilate spike train data from rat layer V cortical neurons into a biophysical model [50], and to assimilate noisy, modelgenerated data for other states to motivate the use of imaging techniques when available [51].
An approach to the variational problem which tries to uncover the global minima more systematically has been developed [32]. In this framework, comparing to (36), they define for diagonal entries of \(Q^{1}\) that
for \(\alpha>1\) and \(\beta\geq0\). The model term is initialized as relatively small, and over the course of an annealing procedure, β is incremented resulting in a steady increase of the model term’s influence on the cost function. This annealing schedule is conducted in parallel for different initial guesses for the statespace. The development of this variational approach can be found in [52], and it has been used to assimilate neuronal data from HVC neurons [34] as well as to calibrate a neuromorphic very large scale integrated (VLSI) circuit [53]. An alternative to the variational approach is to frame the assimilation problem from a probabilistic sampling perspective and use Markov chain MonteCarlo methods [54].
A closely associated variational technique, known as “nudging”, augments the vector field with a control term. If we only have observations of the voltage, this manifests as follows:
The vector field with the observational coupling term is now passed into the strong 4DVar constraints. The control parameter u may remain fixed, or be estimated along with the states [55, 56]. More details on nudging can be found [57]. A similar control framework has been applied to data from neurons of the stomatogastric ganglion [58].
Many other approaches outside the framework of data assimilation have been developed for parameter estimation of neuronal models, see [59] for a review. A problem often encountered when fitting models to a voltage trace is that phase shifts, or small differences in spike timing, between model output and the data can result in large root mean square error. This is less of an issue for data assimilation methods, especially sequential algorithms like UKF. Other approaches to avoid harshly penalizing spike timing errors in the cost function are to consider spikes in the data and modelgenerated spikes that occur within a narrow time window of each other as coincident [60], or to minimize error with respect to the \(dV/dt\) versus V phase–plane trajectory rather than \(V(t)\) itself [59]. Another way to avoid spike mismatch errors is to force the model with the voltage data and perform linear regression to estimate the linear parameters (maximal conductances), and then perhaps couple the problem with another optimization strategy to access the nonlinearlydependent gating parameters [3, 61, 62].
A common optimization strategy is to construct an objective function that encapsulates important features derived from the voltage trace, and then use a genetic algorithm to stochastically search for optimal solutions. These algorithms proceed by forming a population of possible solutions and applying biologically inspired evolution strategies to gradually increase the fitness (defined with respect to the objective function) of the population across generations. Multiobjective optimization schemes will generate a “Pareto front” of optimal solutions that are considered equally good. A multiobjective nondominated sorting genetic algorithm (NSGAII) has recently been used to estimate parameters of the pacemaker PD neurons of the crab pyloric network [63, 64].
In this paper, we compared the bifurcation structure of models estimated by DA algorithms to the bifurcation structure of the model that generated the data. We found that the estimated models exhibited the correct bifurcations even when the algorithms were initiated in a region of parameter space corresponding to a different bifurcation regime. This type of twin experiment is a useful addition to the field that specifically emphasizes the difficulty of nonlinear estimation and provides a qualitative measure of estimation success or failure. Prior literature on parameter estimation that has made use of geometric structure includes work on bursting respiratory neurons [65] and “inverse bifurcation analysis” of gene regulatory networks [66, 67].
Looking forward, data assimilation can complement the growth of new recording technologies for collecting observational data from the brain. The joint collaboration of these automated algorithms with the painstaking work of experimentalists and model developers may help answer many remaining questions about neuronal dynamics.
Abbreviations
 DA:

data assimilation
 PDE:

partial differential equation
 4DVar:

4DVariational
 EKF:

Extended Kalman Filter
 UKF:

Unscented Kalman Filter
 SNIC:

saddlenode on invariant circle
 EnKF:

Ensemble Kalman Filter
 LETK:

Local Ensemble Transform Kalman Filter
References
 1.
Hodgkin AL, Huxley AF. A quantitative description of membrane current and its application to conduction and excitation in nerve. Bull Math Biol. 1990;52(1–2):25–71.
 2.
Meliza CD, Kostuk M, Huang H, Nogaret A, Margoliash D, Abarbanel HD. Estimating parameters and predicting membrane voltages with conductancebased neuron models. Biol Cybern. 2014;108:495–516.
 3.
Lepora NF, Overton PG, Gurney K. Efficient fitting of conductancebased model neurons from somatic current clamp. J Comput Neurosci. 2012;32(1):1–24.
 4.
Carton JA, Giese BS. A reanalysis of ocean climate using simple ocean data assimilation (SODA). Mon Weather Rev. 2008;136(8):2999–3017. https://doi.org/10.1175/2007MWR1978.1.
 5.
Greybush SJ, Wilson RJ, Hoffman RN, Hoffman MJ, Miyoshi T, Ide K, et al.. Ensemble Kalman filter data assimilation of thermal emission spectrometer temperature retrievals into a mars GCM. J Geophys Res, Planets. 2012;117(11):1–17.
 6.
Reichle RH. Data assimilation methods in the Earth sciences. Adv Water Resour. 2008;31(11):1411–8. https://doi.org/10.1016/j.advwatres.2008.01.001.
 7.
Bauer P, Thorpe A, Brunet G. The quiet revolution of numerical weather prediction. Nature. 2015;525(7567):47–55.
 8.
Hoffman MJ, LaVigne NS, Scorse ST, Fenton FH, Cherry EM. Reconstructing threedimensional reentrant cardiac electrical wave dynamics using data assimilation. Chaos, Interdiscip J Nonlinear Sci. 2016;26(1):013107. https://doi.org/10.1063/1.4940238.
 9.
Apte A. An Introduction to data assimilation. In: Sarkar S, Basu U, De S, editors. Applied mathematics. Springer proceedings in mathematics & statistics. 2015. p. 31–42.
 10.
Asch M, Bocquet M, Nodet M. Data assimilation: methods, algorithms, and applications. Fundamentals of algorithms. Philadelphia: SIAM; 2016. https://doi.org/10.1137/1.9781611974546
 11.
Julier SJ, Uhlmann JK. Unscented filtering and nonlinear estimation. In: Proceedings of the IEEE. vol. 92. 2004. p. 401–22. https://doi.org/10.1109/JPROC.2003.823141.
 12.
Gordon NJ, Salmond DJ, Smith AFM. Novel approach to nonlinear/nonGaussian Bayesian state estimation. IEE Proc F, Commun Radar Signal Process. 1993;140(2):107–13.
 13.
Liu J, West M. Combined parameter and state estimation in simulationbased filtering. In: Doucet A, Freitas N, Gordon N, editors. Sequential Monte Carlo methods in practice. Statistics for engineering and information science. New York: Springer; 2001. p. 197–223.
 14.
DelSole T, Yang X. State and parameter estimation in stochastic dynamical models. Physica D. 2010;239(18):1781–8.
 15.
Ullah G, Schiff SJ. Assimilating seizure dynamics. PLoS Comput Biol. 2010;6(5):e1000776. https://doi.org/10.1371/journal.pcbi.1000776.
 16.
Berry T, Sauer T. Adaptive ensemble Kalman filtering of nonlinear systems. Tellus, Ser A Dyn Meteorol Oceanogr. 2013;65(1):2031. https://doi.org/10.3402/tellusa.v65i0.20331
 17.
SedighSarvestani M, Schiff SJ, Gluckman BJ. Reconstructing mammalian sleep dynamics with data assimilation. PLoS Comput Biol. 2012;8(11):e1002788. https://doi.org/10.1371/journal.pcbi.1002788.
 18.
Van der Merwe R, Wan EA. The squareroot unscented Kalman filter for state and parameterestimation. In: 2001 IEEE international conference on acoustics, speech, and signal processing. Proceedings (Cat. No. 01CH37221). vol. 6. New York: IEEE Press; 2001. p. 3461–4. https://doi.org/10.1109/ICASSP.2001.940586
 19.
Rotstein HG, Olarinre M, Golowasch J. Dynamic compensation mechanism gives rise to period and dutycycle level sets in oscillatory neuronal models. J Neurophysiol. 2016;116(5):2431–52. https://doi.org/10.1152/jn.00357.2016
 20.
Walch OJ, Eisenberg MC. Parameter identifiability and identifiable combinations in generalized Hodgkin–Huxley models. Neurocomputing. 2016;199:137–43.
 21.
Stanhope S, Rubin JE, Swigon D. Identifiability of linear and linearinparameters dynamical systems from a single trajectory. SIAM J Appl Dyn Syst. 2014;13(4):1792–815. https://doi.org/10.1137/130937913.
 22.
Stanhope S, Rubin JE, Swigon D. Robustness of solutions of the inverse problem for linear dynamical systems with uncertain data. SIAM/ASA J Uncertain Quantificat. 2017;5(1):572–97. https://doi.org/10.1137/16M1062466.
 23.
Ermentrout GB, Terman DH. Mathematical foundations of neuroscience. Interdisciplinary applied mathematics. vol. 35. New York: Springer; 2010. https://doi.org/10.1007/9780387877082.
 24.
Izhikevich E. Dynamical systems in neuroscience: the geometry of excitability and bursting. Cambridge: MIT Press; 2007.
 25.
Morris C, Lecar H. Voltage oscillations in the barnacle giant muscle fiber. Biophys J. 1981;35(1):193–213.
 26.
Wan EA, Van Der Merwe R. The unscented Kalman filter for nonlinear estimation. In: Proceedings of the IEEE 2000 adaptive systems for signal processing, communications, and control symposium, ASSPCC 2000 (Cat. No. 00EX373). 2000. p. 153–8. https://doi.org/10.1109/ASSPCC.2000.882463
 27.
Byrd RH, Hribar ME, Nocedal J. An interior point algorithm for largescale nonlinear programming. SIAM J Optim. 1999;9(4):877–900.
 28.
Byrd RH, Gilbert JC, Nocedal J. A trust region method based on interior point techniques for nonlinear programming. Math Program, Ser A. 2000;89:149–85.
 29.
Waltz RA, Morales JL, Nocedal J, Orban D. An interior algorithm for nonlinear optimization that combines line search and trust region steps. Math Program, Ser A. 2006;107:391–408.
 30.
Weinstein MJ, Rao AV. Algorithm 984: ADiGator, a toolbox for the algorithmic differentiation of mathematical functions in MATLAB using source transformation via operator overloading. ACM Trans Math Softw. 2017;44(2):1–25. https://doi.org/10.1145/3104990
 31.
Kelly M. An introduction to trajectory optimization: how to do your own direct collocation. SIAM Rev. 2017;59(4):849–904.
 32.
Ye J, Rey D, Kadakia N, Eldridge M, Morone UI, Rozdeba P, et al.. Systematic variational method for statistical nonlinear state and parameter estimation. Phys Rev E, Stat Nonlinear Soft Matter Phys. 2015;92(5):052901. https://doi.org/10.1103/PhysRevE.92.052901.
 33.
Rinzel J, Ermentrout GB. Analysis of neural excitability and oscillations. In: Koch C, Segev I, editors. Methods in neuronal modeling: from synapses to networks. Cambridge: MIT Press; 1989. p. 135–69.
 34.
Kadakia N, Armstrong E, Breen D, Morone U, Daou A, Margoliash D, et al.. Nonlinear statistical data assimilation for HVC\(_{\textrm {RA}}\) neurons in the avian song system. Biol Cybern. 2016;110(6):417–34. https://doi.org/10.1007/s0042201606973.
 35.
Zhu C, Byrd RH, Lu P, Nocedal J. Algorithm 778: LBFGSB: Fortran subroutines for largescale boundconstrained optimization. ACM Trans Math Softw. 1997;23(4):550–60.
 36.
Nocedal J. OPTI Toolbox: a free MATLAB toolbox for optimization. 2018. https://www.inverseproblem.co.nz/OPTI/index.php/Solvers/LBFGSB. Accessed 20180626.
 37.
Becker S. LBFGSBC. 2018. https://github.com/stephenbeckr/LBFGSBC. Accessed 20180626.
 38.
Olufsen MS, Ottesen JT. A practical approach to parameter estimation applied to model predicting heart rate regulation. J Math Biol. 2013;67(1):39–68. https://doi.org/10.1007/s0028501205358.
 39.
Abarbanel HDI, Shirman S, Breen D, Kadakia N, Rey D, Armstrong E, Margoliash D. A unifying view of synchronization for data assimilation in complex nonlinear networks. Chaos. 2017;27(12):126802. https://doi.org/10.1063/1.5001816.
 40.
Law K, Stuart A, Zygalakis K. Data assimilation: a mathematical introduction. In texts in applied mathematics. vol. 62. Cham: Springer; 2015. https://doi.org/10.1007/9783319203256.
 41.
Voss HU, Timmer J, Kurths J. Nonlinear dynamical system identification from uncertain and indirect measurements. Int J Bifurc Chaos. 2004;14(06):1905–33. https://doi.org/10.1142/S0218127404010345.
 42.
Schiff SJ. Neural control engineering: the emerging intersection between control theory and neuroscience. In: Computational neuroscience. Cambridge: MIT Press; 2012. https://books.google.com/books?id=P9UvTQtnqKwC&pgis=1.
 43.
Wei Y, Ullah G, Parekh R, Ziburkus J, Schiff SJ. Kalman filter tracking of intracellular neuronal voltage and current. In: Proceedings of the IEEE conference on decision and control. 2011. p. 5844–9.
 44.
Hamilton F, Berry T, Peixoto N, Sauer T. Realtime tracking of neuronal network structure using data assimilation. Physical Review E. 2013;88(5):052715. https://doi.org/10.1103/PhysRevE.88.052715.
 45.
Lankarany M, Zhu WP, Swamy MNS. Joint estimation of states and parameters of HodgkinHuxley neuronal model using Kalman filtering. Neurocomputing. 2014;136:289–99. https://doi.org/10.1016/j.neucom.2014.01.003.
 46.
Evensen G. The ensemble Kalman filter for combined state and parameter estimation. IEEE Control Syst Mag. 2009;29(3):83–104. http://ieeexplore.ieee.org/document/4939313/.
 47.
Evensen G. Data assimilation: the ensemble Kalman filter. Berlin: Springer; 2009. https://doi.org/10.1007/9783642037115.
 48.
Hunt BR, Kostelich EJ, Szunyogh I. Efficient data assimilation for spatiotemporal chaos: a local ensemble transform Kalman filter. Phys D: Nonlinear Phenom. 2007;230(1–2):112–26.
 49.
Ditlevsen S, Samson A. Estimation in the partially observed stochastic Morris–Lecar neuronal model with particle filter and stochastic approximation methods. Ann Appl Stat. 2014;8(2):674–702. http://projecteuclid.org/euclid.aoas/1404229510.
 50.
Meng L, Kramer MA, Middleton SJ, Whittington MA, Eden UT. A unified approach to linking experimental, statistical and computational analysis of spike train data. PLoS ONE. 2014;9(1):e85269. https://doi.org/10.1371/journal.pone.0085269.
 51.
Huys QJM, Paninski L. Smoothing of, and parameter estimation from, noisy biophysical recordings. PLoS Comput Biol. 2009;5(5):e1000379. https://doi.org/10.1371/journal.pcbi.1000379.
 52.
Abarbanel H. Predicting the future: completing models of observed complex systems. Understanding complex systems. New York: Springer; 2013. https://books.google.com/books?id=Vne5BQAAQBAJ.
 53.
Wang J, Breen D, Akinin A, Broccard F, Abarbanel HDI, Cauwenberghs G. Assimilation of biophysical neuronal dynamics in neuromorphic VLSI. IEEE Trans Biomed Circuits Syst. 2017;11(6):1258–70.
 54.
Kostuk M, Toth BA, Meliza CD, Margoliash D, Abarbanel HDI. Dynamical estimation of neuron and network properties II: path integral Monte Carlo methods. Biol Cybern. 2012;106(3):155–67.
 55.
Toth BA, Kostuk M, Meliza CD, Margoliash D, Abarbanel HDI. Dynamical estimation of neuron and network properties I: variational methods. Biol Cybern. 2011;105(3–4):217–37. https://doi.org/10.1007/s0042201104591.
 56.
Abarbanel HDI, Creveling DR, Farsian R, Kostuk M. Dynamical state and parameter estimation. SIAM J Appl Dyn Syst. 2009;8(4):1341–81. https://doi.org/10.1137/090749761.
 57.
Lakshmivarahan S, Lewis JM. Nudging methods: a critical overview. In: Park SK, Xu L, editors. Data assimilation for atmospheric, oceanic and hydrologic applications, vol. II. Berlin: Springer; 2013. p. 27–57. https://doi.org/10.1007/9783642350887_2.
 58.
Brookings T, Goeritz ML, Marder E. Automatic parameter estimation of multicompartmental neuron models via minimization of trace error with control adjustment. J Neurophysiol. 2014;112(9):2332–48. https://doi.org/10.1152/jn.00007.2014.
 59.
Van Geit W, De Schutter E, Achard P. Automated neuron model optimization techniques: a review. Biol Cybern. 2008;99(4–5):241–51.
 60.
Rossant C. Automatic fitting of spiking neuron models to electrophysiological recordings. Front Neuroinformatics. 2010;4:2. https://doi.org/10.3389/neuro.11.002.2010.
 61.
Huys QJM. Efficient estimation of detailed singleneuron models. J Neurophysiol. 2006;96(2):872–90. https://doi.org/10.1152/jn.00079.2006.
 62.
Haufler D, Morin F, Lacaille JC, Skinner FK. Parameter estimation in singlecompartment neuron models using a synchronizationbased method. Neurocomputing. 2007;70(10–12):1605–10.
 63.
Deb K, Pratap A, Agarwal S, Meyarivan T. A fast and elitist multiobjective genetic algorithm: NSGAII. IEEE Trans Evol Comput. 2002;6(2):182–97.
 64.
Fox DM, Tseng HA, Smolinski TG, Rotstein HG, Nadim F. Mechanisms of generation of membrane potential resonance in a neuron with multiple resonant ionic currents. PLoS Comput Biol. 2017;13(6):1–30.
 65.
Tien JH, Guckenheimer J. Parameter estimation for bursting neural models. J Comput Neurosci. 2006;24:359–73.
 66.
Lu J, Engl HW, Schuster P. Inverse bifurcation analysis: application to simple gene systems. Algorithms Mol Biol. 2006;1:11. https://doi.org/10.1186/17487188111.
 67.
Engl HW, Flamm C, Kügler P, Lu J, Müller S, Schuster P. Inverse problems in systems biology. Inverse Probl. 2009;25:1–51.
Acknowledgements
We thank Tyrus Berry and Franz Hamilton for helpful discussions about the UKF and for sharing code, and Nirag Kadakia and Paul Rozdeba for helpful discussions about 4DVar methods and for sharing code. MM also benefited from lectures and discussions at the Mathematics and Climate Summer Graduate Program held at the University of Kansas in 2016, which was sponsored by the Institute for Mathematics and its Applications and the Mathematics and Climate Research Network.
Availability of data and materials
The MATLAB code used in this study is provided as Supplementary Material.
Funding
This work was supported in part by NSF grants DMS1412877 and DMS1555237, and U.S. Army Research Office grant W911NF1610584. The funding bodies had no role in the design of the study and collection, analysis, and interpretation of data and in writing the manuscript should be declared.
Author information
Affiliations
Contributions
MM wrote the computer code implementing the data assimilation algorithms. MM and CD conceived of the study, performed simulations and analysis, wrote the manuscript, and read and approved the final version of the manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Consent for publication
Not applicable.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Electronic Supplementary Material
Below is the link to the electronic supplementary material.
Appendix
Appendix
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Moye, M.J., Diekman, C.O. Data Assimilation Methods for Neuronal State and Parameter Estimation. J. Math. Neurosc. 8, 11 (2018). https://doi.org/10.1186/s1340801800668
Received:
Accepted:
Published:
Keywords
 Data assimilation
 Neuronal excitability
 Conductancebased models
 Parameter estimation